# my global config global: scrape_interval: 120s # By default, scrape targets every 15 seconds. evaluation_interval: 120s # By default, scrape targets every 15 seconds. # scrape_timeout is set to the global default (10s). # Attach these labels to any time series or alerts when communicating with # external systems (federation, remote storage, Alertmanager). external_labels: monitor: 'my-project' # Load and evaluate rules in this file every 'evaluation_interval' seconds. rule_files: # - "alert.rules" # - "first.rules" # - "second.rules" # A scrape configuration containing exactly one endpoint to scrape: # Here it's Prometheus itself. scrape_configs: # The job name is added as a label `job=` to any timeseries scraped from this config. - job_name: 'prometheus' # Override the global default and scrape targets from this job every 5 seconds. scrape_interval: 5s # metrics_path defaults to '/metrics' # scheme defaults to 'http'. static_configs: - targets: ['mikrotik_exporter:9436', 'mktxp:49090'] # This job queries the Synology Exporter - job_name: 'synology-snmp' static_configs: - targets: - 192.168.0.114 # Actual IP of the Synology NAS # Non default metrics path # Full URL looks something like: http://:9116/snmp?module=synology&target= metrics_path: /snmp params: module: [synology] # Tell the exporter to use the synology module relabel_configs: - source_labels: [__address__] target_label: __param_target - source_labels: [__param_target] target_label: instance - target_label: __address__ replacement: synology_snmp:9116 # SNMP exporter's hostname:port.