Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions asap-dropin/.env
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
# - Docker Desktop (Mac/Windows): http://host.docker.internal:9090
# - Linux (host networking): http://172.17.0.1:9090 (default Docker bridge gateway)
# - Prometheus in another compose: use a shared Docker network and the service name
PROMETHEUS_URL=http://host.docker.internal:9090
PROMETHEUS_URL=http://prometheus:9090

# Scrape interval configured in your Prometheus (in seconds).
PROMETHEUS_SCRAPE_INTERVAL=15
Expand All @@ -12,7 +12,7 @@ PROMETHEUS_SCRAPE_INTERVAL=15
# Port on the host where the remote-write receiver listens.
# Add this to your prometheus.yml:
# remote_write:
# - url: http://localhost:${REMOTE_WRITE_PORT}/receive
# - url: http://localhost:${REMOTE_WRITE_PORT}/api/v1/write
REMOTE_WRITE_PORT=9091

# Port on the host where the ASAPQuery query engine listens.
Expand Down
6 changes: 3 additions & 3 deletions asap-dropin/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -65,15 +65,15 @@ Prometheus needs to send all ingested samples to ASAPQuery so it can build sketc

```yaml
remote_write:
- url: http://localhost:9091/receive
- url: http://localhost:9091/api/v1/write
queue_config:
batch_send_deadline: 1s
sample_age_limit: 5m
```

> **Finding the right `remote_write` URL:** The URL is from Prometheus's perspective.
> - **Prometheus on the same host as Docker:** `http://localhost:9091/receive` (default above)
> - **Prometheus in Docker on the same host:** `http://host.docker.internal:9091/receive` (Mac/Windows) or `http://172.17.0.1:9091/receive` (Linux)
> - **Prometheus on the same host as Docker:** `http://localhost:9091/api/v1/write` (default above)
> - **Prometheus in Docker on the same host:** `http://host.docker.internal:9091/api/v1/write` (Mac/Windows) or `http://172.17.0.1:9091/api/v1/write` (Linux)
> - Change `9091` if you set a different `REMOTE_WRITE_PORT` in `.env`

**Reload Prometheus to apply the change:**
Expand Down
10 changes: 5 additions & 5 deletions asap-query-engine/src/main.rs
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ use query_engine_rust::data_model::QueryLanguage;
use std::fs;
use std::sync::{Arc, RwLock};
use tokio::signal;
use tracing::{error, info, warn};
use tracing::{debug, error, info, warn};

use sketch_core::config::{self, ImplMode};

Expand Down Expand Up @@ -570,14 +570,14 @@ async fn spawn_memory_diagnostics(

// 1. Store diagnostics
let store_diag = store.diagnostic_info();
info!(
debug!(
"[MEMORY_DIAG] Store: {} aggregation(s), {} total time_map entries, {:.2} KB total sketch bytes",
store_diag.num_aggregations,
store_diag.total_time_map_entries,
store_diag.total_sketch_bytes as f64 / 1024.0,
);
for agg in &store_diag.per_aggregation {
info!(
debug!(
"[MEMORY_DIAG] agg_id={}: time_map_len={}, read_counts_len={}, aggregate_objects={}, sketch_bytes={:.2} KB",
agg.aggregation_id,
agg.time_map_len,
Expand All @@ -594,13 +594,13 @@ async fn spawn_memory_diagnostics(
.iter()
.map(|c| c.load(Ordering::Relaxed))
.sum();
info!(
debug!(
"[MEMORY_DIAG] PrecomputeEngine: {} total groups across {} workers",
total_groups,
diag.worker_group_counts.len(),
);
for (i, counter) in diag.worker_group_counts.iter().enumerate() {
info!(
debug!(
"[MEMORY_DIAG] worker_{}: group_states_len={}",
i,
counter.load(Ordering::Relaxed),
Expand Down
2 changes: 1 addition & 1 deletion asap-query-engine/src/stores/simple_map_store/per_key.rs
Original file line number Diff line number Diff line change
Expand Up @@ -499,7 +499,7 @@ impl Store for SimpleMapStorePerKey {
let store_data_lock = match self.store.get(&store_key) {
Some(lock) => lock,
None => {
info!("Metric {} not found in store", metric);
debug!("Metric {} not found in store", metric);
return Ok(HashMap::new());
}
};
Expand Down
Loading