@@ -7,12 +7,13 @@ logfile=/var/log/supervisord.log
7
7
# Cannot place this in Celery for now because Celery must run as a single process (see note below)
8
8
# Indexing uses multi-processing to speed things up
9
9
[program:document_indexing]
10
- environment=CURRENT_PROCESS_IS_AN_INDEXING_JOB=true,LOG_FILE_NAME=document_indexing
10
+ environment=CURRENT_PROCESS_IS_AN_INDEXING_JOB=true
11
11
command=python danswer/background/update.py
12
+ stdout_logfile=/var/log/document_indexing.log
13
+ stdout_logfile_maxbytes=16MB
12
14
redirect_stderr=true
13
15
autorestart=true
14
16
15
-
16
17
# Background jobs that must be run async due to long time to completion
17
18
# NOTE: due to an issue with Celery + SQLAlchemy
18
19
# (https://github.com/celery/celery/issues/7007#issuecomment-1740139367)
@@ -87,7 +88,8 @@ stopasgroup=true
87
88
# More details on setup here: https://docs.danswer.dev/slack_bot_setup
88
89
[program:slack_bot]
89
90
command=python danswer/danswerbot/slack/listener.py
90
- environment=LOG_FILE_NAME=slack_bot
91
+ stdout_logfile=/var/log/slack_bot.log
92
+ stdout_logfile_maxbytes=16MB
91
93
redirect_stderr=true
92
94
autorestart=true
93
95
startretries=5
@@ -101,8 +103,8 @@ command=tail -qF
101
103
/var/log/celery_worker_primary.log
102
104
/var/log/celery_worker_light.log
103
105
/var/log/celery_worker_heavy.log
104
- /var/log/document_indexing_info .log
105
- /var/log/slack_bot_debug .log
106
+ /var/log/document_indexing .log
107
+ /var/log/slack_bot .log
106
108
stdout_logfile=/dev/stdout
107
109
stdout_logfile_maxbytes = 0 # must be set to 0 when stdout_logfile=/dev/stdout
108
110
autorestart=true
0 commit comments