@@ -73,11 +73,13 @@ class FindDoppler:
73
73
Append output DAT & LOG files? (True/False)
74
74
log_level_int : int, optional
75
75
Python logging threshold level (INFO, DEBUG, or WARNING)
76
+ blank_dc : bool, optional
77
+ Use blank_dc() for spike smoothing.
76
78
77
79
"""
78
80
def __init__ (self , datafile , max_drift = 4.0 , min_drift = 0.00001 , snr = 25.0 , out_dir = './' , coarse_chans = None ,
79
81
obs_info = None , flagging = False , n_coarse_chan = None , kernels = None , gpu_backend = False , gpu_id = 0 ,
80
- precision = 2 , append_output = False , log_level_int = logging .INFO ):
82
+ precision = 2 , append_output = False , log_level_int = logging .INFO , blank_dc = True ):
81
83
82
84
print (version_announcements )
83
85
@@ -113,10 +115,11 @@ def __init__(self, datafile, max_drift=4.0, min_drift=0.00001, snr=25.0, out_dir
113
115
self .status = True
114
116
self .flagging = flagging
115
117
self .append_output = append_output
118
+ self .flag_blank_dc = blank_dc
116
119
self .parms = 'datafile={}, max_drift={}, min_drift={}, snr={}, out_dir={}, coarse_chans={}' \
117
120
.format (datafile , max_drift , min_drift , snr , out_dir , coarse_chans ) \
118
- + ', flagging={}, n_coarse_chan={}, kernels={}, gpu_id={}, gpu_backend={}' \
119
- .format (flagging , self .n_coarse_chan , kernels , gpu_id , gpu_backend ) \
121
+ + ', flagging={}, n_coarse_chan={}, kernels={}, gpu_id={}, gpu_backend={}, blank_dc={} ' \
122
+ .format (flagging , self .n_coarse_chan , kernels , gpu_id , gpu_backend , blank_dc ) \
120
123
+ ', precision={}, append_output={}, log_level_int={}, obs_info={}' \
121
124
.format (precision , append_output , log_level_int , obs_info )
122
125
if min_drift < 0 or max_drift < 0 :
@@ -166,7 +169,6 @@ def search(self, n_partitions=1, progress_bar='n'):
166
169
167
170
# As of 2.1.0, add max_drift_rate and obs_length to FileWriter header input
168
171
header_in ['max_drift_rate' ] = self .max_drift
169
- #header_in['obs_length'] was already set in data_handler.py DATAH __init__
170
172
171
173
wfilename = filename_in .split ('/' )[- 1 ].replace ('.h5' , '' ).replace ('.fits' , '' ).replace ('.fil' , '' )
172
174
path_log = '{}/{}.log' .format (self .out_dir .rstrip ('/' ), wfilename )
@@ -198,7 +200,7 @@ def search(self, n_partitions=1, progress_bar='n'):
198
200
199
201
# Run serial version
200
202
if n_partitions == 1 :
201
- sched = Scheduler (load_the_data , [ (l , self .kernels .precision ) for l in self .data_handle .data_list ])
203
+ sched = Scheduler (load_the_data , [ (l , self .kernels .precision , self . flag_blank_dc ) for l in self .data_handle .data_list ])
202
204
for dl in self .data_handle .data_list :
203
205
search_coarse_channel (dl , self , dataloader = sched , filewriter = filewriter , logwriter = logwriter )
204
206
# Run Parallel version via dask
@@ -216,15 +218,15 @@ def search(self, n_partitions=1, progress_bar='n'):
216
218
t1 = time .time ()
217
219
self .last_logwriter (path_log , '\n ===== Search time: {:.2f} minutes' .format ((t1 - t0 )/ 60.0 ))
218
220
219
- def load_the_data (d , precision ):
221
+ def load_the_data (d , precision , flag_blank_dc ):
220
222
data_obj = DATAH5 (d ['filename' ],
221
223
f_start = d ['f_start' ],
222
224
f_stop = d ['f_stop' ],
223
225
coarse_chan = d ['coarse_chan' ],
224
226
n_coarse_chan = d ['n_coarse_chan' ],
225
227
gpu_backend = False ,
226
228
precision = precision )
227
- spectra , drift_indices = data_obj .load_data ()
229
+ spectra , drift_indices = data_obj .load_data (flag_blank_dc = flag_blank_dc )
228
230
data_obj .close ()
229
231
230
232
return (data_obj , spectra , drift_indices )
@@ -277,7 +279,7 @@ def search_coarse_channel(data_dict, find_doppler_instance, dataloader=None, log
277
279
if dataloader :
278
280
data_obj , spectra , drift_indices = dataloader .get ()
279
281
else :
280
- data_obj , spectra , drift_indices = load_the_data (d , fd .kernels .precision )
282
+ data_obj , spectra , drift_indices = load_the_data (d , fd .kernels .precision , fd . flag_blank_dc )
281
283
282
284
fileroot_out = filename_in .split ('/' )[- 1 ].replace ('.h5' , '' ).replace ('.fits' , '' ).replace ('.fil' , '' )
283
285
if logwriter is None :
0 commit comments