@@ -261,7 +261,7 @@ def _run_indexing(
261
261
3. Updates Postgres to record the indexed documents + the outcome of this run
262
262
"""
263
263
start_time = time .monotonic () # jsut used for logging
264
-
264
+ logger . error ( "Starting indexing run" )
265
265
with get_session_with_current_tenant () as db_session_temp :
266
266
index_attempt_start = get_index_attempt (db_session_temp , index_attempt_id )
267
267
if not index_attempt_start :
@@ -315,6 +315,7 @@ def _run_indexing(
315
315
# don't go into "negative" time if we've never indexed before
316
316
window_start = datetime .fromtimestamp (0 , tz = timezone .utc )
317
317
318
+ logger .error ("Getting most recent attempt" )
318
319
most_recent_attempt = next (
319
320
iter (
320
321
get_recent_completed_attempts_for_cc_pair (
@@ -326,6 +327,7 @@ def _run_indexing(
326
327
),
327
328
None ,
328
329
)
330
+ logger .error (f"Most recent attempt: { most_recent_attempt } " )
329
331
# if the last attempt failed, try and use the same window. This is necessary
330
332
# to ensure correctness with checkpointing. If we don't do this, things like
331
333
# new slack channels could be missed (since existing slack channels are
@@ -361,6 +363,7 @@ def _run_indexing(
361
363
httpx_client = HttpxPool .get ("vespa" ),
362
364
)
363
365
366
+ logger .error ("Building indexing pipeline" )
364
367
indexing_pipeline = build_indexing_pipeline (
365
368
embedder = embedding_model ,
366
369
information_content_classification_model = information_content_classification_model ,
@@ -782,7 +785,7 @@ def run_indexing_entrypoint(
782
785
callback : IndexingHeartbeatInterface | None = None ,
783
786
) -> None :
784
787
"""Don't swallow exceptions here ... propagate them up."""
785
-
788
+ logger . error ( "Starting indexing run: run_indexing_entrypoint" )
786
789
if is_ee :
787
790
global_version .set_ee ()
788
791
0 commit comments