Commit dc5f7d28 authored by Kipp Cannon's avatar Kipp Cannon

gstlal, gstlal-ugly, gstlal-inspiral:

- a variety of import related adjustments and fixes
parent c99bfe59
......@@ -27,11 +27,11 @@ from glue.ligolw import ligolw
from glue.ligolw import lsctables
from glue.ligolw import utils as ligolw_utils
from glue.ligolw.utils import process as ligolw_process
from lal import MSUN_SI
from glue import lal as gluelal
from gstlal import templates
from gstlal import inspiral_pipe
from gstlal import chirptime
import lal
from lal.utils import CacheEntry
## @file gstlal_bank_splitter
#
......@@ -199,7 +199,7 @@ for bank_count, (_, rows) in enumerate(outputrows):
for row in rows:
# Chirptime uses SI
m1_SI, m2_SI = MSUN_SI * row.mass1, MSUN_SI * row.mass2
m1_SI, m2_SI = lal.MSUN_SI * row.mass1, lal.MSUN_SI * row.mass2
# Find the total spin magnitudes
spin1, spin2 = (row.spin1x**2 + row.spin1y**2 + row.spin1z**2)**.5, (row.spin2x**2 + row.spin2y**2 + row.spin2z**2)**.5
......@@ -226,7 +226,7 @@ for bank_count, (_, rows) in enumerate(outputrows):
sngl_inspiral_table[:] = rows
output = inspiral_pipe.T050017_filename(options.instrument, "%04d_GSTLAL_SPLIT_BANK" % bank_count, 0, 0, ".xml.gz", path = options.output_path)
output_cache_file.write("%s\n" % gluelal.CacheEntry.from_T050017("file://localhost%s" % os.path.abspath(output)))
output_cache_file.write("%s\n" % CacheEntry.from_T050017("file://localhost%s" % os.path.abspath(output)))
ligolw_utils.write_filename(xmldoc, output, gz = output.endswith('gz'), verbose = options.verbose)
output_cache_file.close()
......@@ -199,8 +199,8 @@ Gst.init(None)
import lal
from lal import LIGOTimeGPS
from lal.utils import CacheEntry
from glue.lal import CacheEntry
from glue import segments
from glue import segmentsUtils
from glue.ligolw import ligolw
......
......@@ -43,7 +43,6 @@ from optparse import OptionParser
import sys
from glue.lal import CacheEntry
from glue.ligolw import ligolw
from glue.ligolw import lsctables
from glue.ligolw import utils as ligolw_utils
......@@ -51,10 +50,11 @@ from glue.ligolw.utils import process as ligolw_process
from glue.ligolw.utils import search_summary as ligolw_search_summary
from glue.ligolw.utils import segments as ligolw_segments
from glue import segments
from pylal import ligolw_burca2
from lalburst import ligolw_burca2
from pylal import ligolw_thinca
from pylal import snglcoinc
from gstlal import far
from lal.utils import CacheEntry
process_name = u"gstlal_inspiral_calc_likelihood"
......
......@@ -40,12 +40,12 @@ from optparse import OptionParser
import sys
from glue.lal import CacheEntry
from glue.ligolw import ligolw
from glue.ligolw import utils as ligolw_utils
from glue.ligolw.utils import process as ligolw_process
from glue.ligolw.utils import search_summary as ligolw_search_summary
from gstlal import far
from lal.utils import CacheEntry
__author__ = "Kipp Cannon <kipp.cannon@ligo.org>"
......
......@@ -41,11 +41,11 @@ from optparse import OptionParser
from glue import segments
from glue.ligolw import ligolw
from glue.ligolw import lsctables
from glue.lal import CacheEntry
from glue.lal import Cache
from glue.ligolw import utils as ligolw_utils
from lal.series import *
import lal
from lal.series import *
from lal.utils import CacheEntry
import lalsimulation
import multiprocessing
......
......@@ -49,11 +49,11 @@ matplotlib.rcParams.update({
import lal
from lal import LIGOTimeGPS
from lal.utils import CacheEntry
from glue.ligolw import ligolw
from glue.ligolw import table
from glue.ligolw import lsctables
from glue.ligolw import utils as ligolw_utils
from glue import lal as gluelal
from gstlal import lvalert_helper
from ligo.gracedb import rest as gracedb
from pylal import rate
......@@ -143,7 +143,7 @@ subprocess.call(["ls %s/* | lalapps_path2cache >> %s/%s.cache" % (str(gps + 1000
# FIXME don't hard code this regex
pattern = re.compile("0[0-9]{3}_LLOID$")
pool = multiprocessing.Pool(8)
ce = [gluelal.CacheEntry(l) for l in open("%s/%s.cache" % (wd, str(gps)))]
ce = [CacheEntry(l) for l in open("%s/%s.cache" % (wd, str(gps)))]
sngl = []
for res in pool.map(parse, [c for c in ce if gps in c.segment and pattern.match(c.description)]):
sngl.extend(res)
......
......@@ -52,7 +52,7 @@ from glue.ligolw import utils as ligolw_utils
from glue.ligolw.utils import process as ligolw_process
from glue.ligolw.utils import search_summary as ligolw_search_summary
from glue import segments
from glue import lal
from lal.utils import CacheEntry
from gstlal import far
......@@ -81,7 +81,7 @@ def parse_command_line():
options, urls = parser.parse_args()
if options.likelihood_cache:
urls += [lal.CacheEntry(line).url for line in open(options.likelihood_cache)]
urls += [CacheEntry(line).url for line in open(options.likelihood_cache)]
if not urls and not options.ignore_missing:
raise ValueError("no input documents")
......
......@@ -87,22 +87,22 @@ __author__ = 'Chad Hanna <chad.hanna@ligo.org>'
##############################################################################
# import standard modules and append the lalapps prefix to the python path
import sys, os, copy, math, stat
import subprocess, socket, tempfile
import sys, os, stat
import itertools
from optparse import OptionParser
##############################################################################
# import the modules we need to build the pipeline
import lal
import lal.series
from lal.utils import CacheEntry
from glue import pipeline
from glue.lal import Cache, CacheEntry
from glue.lal import Cache
from glue import segments
from glue.ligolw import ligolw
from glue.ligolw import lsctables
import glue.ligolw.utils as ligolw_utils
import glue.ligolw.utils.segments as ligolw_segments
from optparse import OptionParser
from gstlal import inspiral, inspiral_pipe
from gstlal import dagparts as gstlaldagparts
from gstlal import datasource
......
......@@ -53,10 +53,10 @@ import sqlite3
import sys
from glue.lal import CacheEntry
from glue.ligolw import dbtables
from glue.ligolw import lsctables
from glue import segmentsUtils
from lal.utils import CacheEntry
from pylal import ligolw_thinca
......
......@@ -42,10 +42,10 @@ from glue.ligolw import dbtables
from glue.ligolw import utils as ligolw_utils
from glue.ligolw import array as ligolw_array
from glue import segmentsUtils
from glue.lal import CacheEntry
from glue import iterutils
import lal
from lal.utils import CacheEntry
from pylal import rate
from pylal import imr_utils
......
......@@ -55,13 +55,12 @@ import sqlite3
import sys
import lal
from lal.utils import CacheEntry
from glue import segments
from glue.ligolw import dbtables
from glue.ligolw import lsctables
from glue.ligolw.utils import segments as ligolw_segments
from glue.lal import CacheEntry
from pylal import git_version
from gstlal import far
from gstlal import inspiral_pipe
from gstlal import gviz_api
......@@ -110,8 +109,8 @@ lsctables.SnglInspiralTable.RowType = SnglInspiral
__author__ = "Kipp Cannon <kipp.cannon@ligo.org>, Chad Hanna <channa@ligo.caltech.edu>"
__version__ = "git id %s" % git_version.id
__date__ = git_version.date
__version__ = "git id %s" % "FIXME"
__date__ = "FIXME"
#
......@@ -125,7 +124,7 @@ __date__ = git_version.date
def parse_command_line():
parser = OptionParser(
version = "Name: %%prog\n%s" % git_version.verbose_msg
version = "Name: %%prog\n%s" % "" # FIXME
)
parser.add_option("","--input-cache", help = "Also get the list of databases to process from this LAL cache.")
parser.add_option("--user-tag", metavar = "user-tag", default = "ALL", help = "Set the prefix for output filenames (default = \"ALL\").")
......@@ -156,7 +155,7 @@ def parse_command_line():
if not filenames:
filenames = []
if options.input_cache:
filenames.extend(c.path for c in map(lal.CacheEntry, open(options.input_cache)))
filenames.extend(c.path for c in map(CacheEntry, open(options.input_cache)))
if options.likelihood_file is None:
raise ValueError("Must provide likelihood file.")
......
......@@ -57,13 +57,13 @@ import sys
from glue import segments
from glue.lal import CacheEntry
from glue.ligolw import ligolw
from glue.ligolw import dbtables
from glue.ligolw import lsctables
from glue.ligolw import utils as ligolw_utils
from glue.ligolw.utils import process as ligolw_process
from glue.text_progress_bar import ProgressBar
from lal.utils import CacheEntry
from pylal import ligolw_thinca
......
......@@ -24,7 +24,6 @@ from optparse import OptionParser
import sqlite3
from time import gmtime
from glue.lal import CacheEntry
from glue.ligolw import dbtables
from glue.ligolw import lsctables
from glue.ligolw import utils as ligolw_utils
......@@ -33,8 +32,9 @@ from glue.offsetvector import offsetvector
from glue import segments
from gstlal import inspiral
from gstlal import far
from lal import UTCToGPS
from pylal import ligolw_burca2
import lal
from lal.utils import CacheEntry
from lalburst import ligolw_burca2
from pylal import snglcoinc
# FIXME Need to add entries to process and process_params tables
......@@ -105,7 +105,7 @@ class Connections:
self.verbose = options.verbose
for n, database in enumerate(databases):
working_filename = dbtables.get_connection_filename(database, tmp_path = options.tmp_space, verbose = self.verbose)
self.connection_id_map['connection_id:%d' % n] = namedtuple('ConnectionNamedTuple', 'database, working_filename, connection, process_start_time')('%s' % database, '%s' % working_filename, sqlite3.connect(working_filename), '%d' % UTCToGPS(gmtime()))
self.connection_id_map['connection_id:%d' % n] = namedtuple('ConnectionNamedTuple', 'database, working_filename, connection, process_start_time')('%s' % database, '%s' % working_filename, sqlite3.connect(working_filename), '%d' % lal.UTCToGPS(gmtime()))
#
# Create dictionaries keyed by the connection_id value set
......
......@@ -23,24 +23,20 @@ This program makes a dag to generate svd banks
__author__ = 'Chad Hanna <chad.hanna@ligo.org>'
#
# import standard modules and append the lalapps prefix to the python path
# import standard modules
#
import sys, os, copy, math
import subprocess, socket, tempfile
import sys, os, math
from optparse import OptionParser
#
# import the modules we need to build the pipeline
#
from glue import iterutils
from glue import pipeline
from glue import lal
from glue.ligolw import lsctables
from glue import segments
from optparse import OptionParser
from gstlal import inspiral_pipe
from gstlal import far
from lal.utils import CacheEntry
## @file gstlal_inspiral_svd_bank_pipe
# This program will make a HTCondor DAG to automate the creation of svd bank files; see gstlal_inspiral_svd_bank_pipe for more information
......@@ -146,7 +142,7 @@ except:
dag = inspiral_pipe.DAG(options.output_name)
svdJob = inspiral_pipe.generic_job("gstlal_svd_bank", tag_base = "gstlal_svd_bank_%s" % ifo, condor_commands = {"request_memory":"1999"})
# Assumes cache is sorted by chirpmass or whatever the SVD sorting algorithm that was chosen
files = [lal.CacheEntry(line).path for line in open(options.bank_cache)]
files = [CacheEntry(line).path for line in open(options.bank_cache)]
#
......@@ -162,7 +158,7 @@ for i, f in enumerate(groups):
bank_ids = range(bank_ids[-1] + 1, bank_ids[-1] + 1 + len(f))
svd_bank_name = inspiral_pipe.T050017_filename(ifo, "GSTLAL_SVD_BANK_%d" % i, 0, 0, ".xml.gz", path = svdJob.output_path)
svd_bank_name = os.path.join(os.getcwd(), svd_bank_name)
dag.output_cache.append(lal.CacheEntry(ifo, "GSTLAL_SVD_BANK_%d" % i, segments.segment(0, 0), "file://localhost%s" % (svd_bank_name,)))
dag.output_cache.append(CacheEntry(ifo, "GSTLAL_SVD_BANK_%d" % i, segments.segment(0, 0), "file://localhost%s" % (svd_bank_name,)))
svdNode = inspiral_pipe.generic_node(svdJob, dag, [],
opts = {"snr-threshold":far.ThincaCoincParamsDistributions.snr_min,
......
......@@ -54,8 +54,8 @@ from glue.ligolw import utils as ligolw_utils
from glue.ligolw.utils import process as ligolw_process
from glue.ligolw import ligolw
from glue import segments
from glue.lal import CacheEntry
from lal import LIGOTimeGPS
from lal.utils import CacheEntry
from pylal import rate
from optparse import OptionParser
......
......@@ -29,34 +29,21 @@ __author__ = 'Chad Hanna <channa@caltech.edu>'
# import standard modules and append the lalapps prefix to the python path
#
import sys, os, copy, stat
import sys, os, stat
import shutil
import socket
from optparse import OptionParser
#
# import the modules we need to build the pipeline
#
from glue import pipeline
from glue import lal
from glue.ligolw import lsctables
from glue.ligolw.ligolw import LIGOLWContentHandler
import glue.ligolw.utils as ligolw_utils
from glue.ligolw.utils import process as ligolw_process
from optparse import OptionParser
from gstlal import inspiral
from gstlal import inspiral_pipe
from gstlal import dagparts
from gstlal import datasource
#
# set up a content handler to open the injection file
#
class ligolwcontenthandler(LIGOLWContentHandler):
pass
lsctables.use_in(ligolwcontenthandler)
from lal.utils import CacheEntry
##
# ### Graph of the HTCondor DAG
......@@ -273,8 +260,8 @@ def parse_command_line():
options.state_vector_on_off_dict = inspiral.state_vector_on_off_dict_from_bit_lists(options.state_vector_on_bits, options.state_vector_off_bits)
options.likelihood_files = [lal.CacheEntry(line).url for line in open(options.likelihood_cache)]
options.zerolag_likelihood_files = [lal.CacheEntry(line).url for line in open(options.zerolag_likelihood_cache)]
options.likelihood_files = [CacheEntry(line).url for line in open(options.likelihood_cache)]
options.zerolag_likelihood_files = [CacheEntry(line).url for line in open(options.zerolag_likelihood_cache)]
return options, filenames, bankcache, channel_dict, dq_channel_dict, framexmit_dict, inj_channel_dict, inj_dq_channel_dict, inj_framexmit_dict, inj_name_dict, inj_range_dict
......
......@@ -35,7 +35,7 @@ import uuid
import lal
import lal.series
from glue.lal import CacheEntry
from lal.utils import CacheEntry
from glue.ligolw import utils as ligolw_utils
from gstlal import svd_bank
from gstlal.far import ThincaCoincParamsDistributions
......
......@@ -55,10 +55,10 @@ from gstlal import simplehandler
class tempcache(list):
"""
List-like object to hold lal.CacheEntry objects, and run
os.unlink() on the .path of each as they are removed from the list
or when the list is garbage collected. All errors during file
removal are ignored.
List-like object to hold CacheEntry objects, and run os.unlink() on
the .path of each as they are removed from the list or when the
list is garbage collected. All errors during file removal are
ignored.
Note that there is no way to remove a CacheEntry from this list
without the file it represents being deleted. If, after adding a
......@@ -69,7 +69,7 @@ class tempcache(list):
Example:
>>> from glue.lal import CacheEntry
>>> from lal.utils import CacheEntry
>>> # create a cache, and add an entry
>>> cache = tempcache()
>>> cache.append(CacheEntry("- - - - file://localhost/tmp/blah.txt"))
......
......@@ -39,8 +39,10 @@
import sys, os
import subprocess, socket, tempfile, copy, doctest
from glue import pipeline, lal
from glue.ligolw import utils, lsctables, array, ligolw
from glue.ligolw import lsctables, ligolw
from glue.ligolw import utils as ligolw_utils
from gstlal import svd_bank
from lal.utils import CacheEntry
#
......@@ -254,7 +256,7 @@ class generic_node(InspiralNode):
for opt, val in input_cache_files.items():
if not os.path.isdir(cache_dir):
os.mkdir(cache_dir)
cache_entries = [lal.CacheEntry.from_T050017("file://localhost%s" % os.path.abspath(filename)) for filename in val]
cache_entries = [CacheEntry.from_T050017("file://localhost%s" % os.path.abspath(filename)) for filename in val]
if input_cache_file_name is None:
cache_file_name = group_T050017_filename_from_T050017_files(cache_entries, '.cache', path = cache_dir)
else:
......@@ -268,7 +270,7 @@ class generic_node(InspiralNode):
for opt, val in output_cache_files.items():
if not os.path.isdir(cache_dir):
os.mkdir(cache_dir)
cache_entries = [lal.CacheEntry.from_T050017("file://localhost%s" % os.path.abspath(filename)) for filename in val]
cache_entries = [CacheEntry.from_T050017("file://localhost%s" % os.path.abspath(filename)) for filename in val]
cache_file_name = group_T050017_filename_from_T050017_files(cache_entries, '.cache', path = cache_dir)
with open(cache_file_name, "w") as cache_file:
lal.Cache(cache_entries).tofile(cache_file)
......@@ -349,7 +351,7 @@ def build_bank_groups(cachedict, numbanks = [2], maxjobs = None):
"""
outstrs = []
ifos = sorted(cachedict.keys())
files = zip(*[[lal.CacheEntry(f).path for f in open(cachedict[ifo],'r').readlines()] for ifo in ifos])
files = zip(*[[CacheEntry(f).path for f in open(cachedict[ifo],'r').readlines()] for ifo in ifos])
for n, bank_group in enumerate(group(files, numbanks)):
if maxjobs is not None and n > maxjobs:
break
......@@ -448,10 +450,10 @@ def group_T050017_filename_from_T050017_files(cache_entries, extension, path = N
def get_svd_bank_params_online(svd_bank_cache):
template_mchirp_dict = {}
for ce in [lal.CacheEntry(f) for f in open(svd_bank_cache)]:
for ce in [CacheEntry(f) for f in open(svd_bank_cache)]:
if not template_mchirp_dict.setdefault("%04d" % int(ce.description.split("_")[3]), []):
min_mchirp, max_mchirp = float("inf"), 0
xmldoc = utils.load_url(ce.path, contenthandler = svd_bank.DefaultContentHandler)
xmldoc = ligolw_utils.load_url(ce.path, contenthandler = svd_bank.DefaultContentHandler)
for root in (elem for elem in xmldoc.getElementsByTagName(ligolw.LIGO_LW.tagName) if elem.hasAttribute(u"Name") and elem.Name == "gstlal_svd_bank_Bank"):
snglinspiraltable = lsctables.SnglInspiralTable.get_table(root)
mchirp_column = snglinspiraltable.getColumnByName("mchirp")
......@@ -465,12 +467,12 @@ def get_svd_bank_params(svd_bank_cache, online = False):
bgbin_file_map = {}
max_time = 0
template_mchirp_dict = {}
for ce in sorted([lal.CacheEntry(f) for f in open(svd_bank_cache)], cmp = lambda x,y: cmp(int(x.description.split("_")[0]), int(y.description.split("_")[0]))):
for ce in sorted([CacheEntry(f) for f in open(svd_bank_cache)], cmp = lambda x,y: cmp(int(x.description.split("_")[0]), int(y.description.split("_")[0]))):
if not online:
bgbin_file_map.setdefault(ce.observatory, []).append(ce.path)
if not template_mchirp_dict.setdefault(ce.description.split("_")[0], []):
min_mchirp, max_mchirp = float("inf"), 0
xmldoc = utils.load_url(ce.path, contenthandler = svd_bank.DefaultContentHandler)
xmldoc = ligolw_utils.load_url(ce.path, contenthandler = svd_bank.DefaultContentHandler)
for root in (elem for elem in xmldoc.getElementsByTagName(ligolw.LIGO_LW.tagName) if elem.hasAttribute(u"Name") and elem.Name == "gstlal_svd_bank_Bank"):
snglinspiraltable = lsctables.SnglInspiralTable.get_table(root)
mchirp_column = snglinspiraltable.getColumnByName("mchirp")
......
#!/usr/bin/python
import sys
from glue import lal
from glue import segments
from glue.ligolw import lsctables
from glue.ligolw import ligolw
from glue.ligolw import utils
from glue.ligolw.utils import segments as ligolw_segments
from glue.ligolw.utils import process as ligolw_process
from lal.utils import CacheEntry
def to_segmentlistdict(l):
s = segments.segmentlistdict()
#FIXME this may not work
key = l.description[:2]
s[key] = segments.segmentlist([l.segment])
return s
return segments.segmentlistdict({
# FIXME this may not work
l.description[:2]: segments.segmentlist([l.segment])
})
segs = segments.segmentlistdict()
for seg in [to_segmentlistdict(lal.CacheEntry(l)) for l in open(sys.argv[1]).readlines()]:
for seg in [to_segmentlistdict(CacheEntry(l)) for l in open(sys.argv[1]).readlines()]:
segs+= seg
xmldoc = ligolw.Document()
......
......@@ -24,21 +24,18 @@ __author__ = 'Chad Hanna <chad.hanna@ligo.org>'
##############################################################################
# import standard modules and append the lalapps prefix to the python path
import sys, os, copy, math
import subprocess, socket, tempfile
# import standard modules
import sys, os, subprocess, socket, tempfile
##############################################################################
# import the modules we need to build the pipeline
from glue import iterutils
from glue import pipeline
from glue import lal
from glue.ligolw import lsctables
from glue import segments
import glue.ligolw.utils as utils
import glue.ligolw.utils as ligolw_utils
import glue.ligolw.utils.segments as ligolw_segments
from optparse import OptionParser
from gstlal import datasource
from lal.utils import CacheEntry
def which(prog):
which = subprocess.Popen(['which',prog], stdout=subprocess.PIPE)
......@@ -164,7 +161,7 @@ class gstlal_reference_psd_node(pipeline.CondorDAGNode):
path = os.getcwd()
output_name = self.output_name = '%s/%s-%d-%d-reference_psd.xml.gz' % (path, instrument, gps_start_time, gps_end_time)
self.add_var_opt("write-psd",output_name)
dag.output_cache.append(lal.CacheEntry(instrument, "-", segments.segment(gps_start_time, gps_end_time), "file://localhost/%s" % (output_name,)))
dag.output_cache.append(CacheEntry(instrument, "-", segments.segment(gps_start_time, gps_end_time), "file://localhost/%s" % (output_name,)))
for p in p_node:
self.add_parent(p)
dag.add_node(self)
......@@ -310,7 +307,7 @@ try: os.mkdir("logs")
except: pass
dag = bank_DAG("recolor_pipe")
seglists = ligolw_segments.segmenttable_get_by_name(utils.load_filename(options.segment_file), "datasegments").coalesce()
seglists = ligolw_segments.segmenttable_get_by_name(ligolw_utils.load_filename(options.segment_file), "datasegments").coalesce()
breakupsegs(seglists, options.min_segment_length)
psdJob = gstlal_reference_psd_job()
......
......@@ -27,27 +27,21 @@ __author__ = 'Chad Hanna <chad.hanna@ligo.org>'
##############################################################################
# import standard modules and append the lalapps prefix to the python path
# import standard modules
import sys, os, copy, math
from optparse import OptionParser
import subprocess, socket, tempfile
##############################################################################
# import the modules we need to build the pipeline
from glue import iterutils
from glue import pipeline
from glue import lal
from glue.ligolw import lsctables
from glue import segments
from glue.ligolw import ligolw
import glue.ligolw.utils as utils
import glue.ligolw.utils as ligolw_utils
import glue.ligolw.utils.segments as ligolw_segments
from optparse import OptionParser
from gstlal import datasource
from gstlal import dagparts
class ContentHandler(ligolw.LIGOLWContentHandler):
pass
lsctables.use_in(ContentHandler)
from lal import series as lalseries
from lal.utils import CacheEntry
#
# Classes for generating reference psds
......@@ -128,7 +122,7 @@ class gstlal_reference_psd_node(pipeline.CondorDAGNode):
path = os.getcwd()
output_name = self.output_name = '%s/%s-%d-%d-reference_psd.xml.gz' % (path, instrument, gps_start_time, gps_end_time)
self.add_var_opt("write-psd",output_name)
dag.output_cache.append(lal.CacheEntry(instrument, "-", segments.segment(gps_start_time, gps_end_time), "file://localhost/%s" % (output_name,)))
dag.output_cache.append(CacheEntry(instrument, "-", segments.segment(gps_start_time, gps_end_time), "file://localhost/%s" % (output_name,)))
for p in p_node:
self.add_parent(p)
dag.add_node(self)
......@@ -285,7 +279,7 @@ except:
dag = dagparts.CondorDAG("gstlal_fake_frames_pipe")
seglists = ligolw_segments.segmenttable_get_by_name(utils.load_filename(options.frame_segments_file, verbose = options.verbose, contenthandler = ContentHandler), options.frame_segments_name).coalesce()
seglists = ligolw_segments.segmenttable_get_by_name(ligolw_utils.load_filename(options.frame_segments_file, verbose = options.verbose, contenthandler = ligolw_segments.LIGOLWContentHandler), options.frame_segments_name).coalesce()
choosesegs(seglists, options.min_segment_length)
psdJob = gstlal_reference_psd_job()
......@@ -316,7 +310,7 @@ if options.whiten_type in ("psdperseg", "medianofpsdperseg"):
psd[instrument] = mediannode[instrument].output_name
elif options.whiten_type is not None:
psd = lalseries.read_psd_xmldoc(utils.load_filename(options.whiten_reference_psd, verbose = options.verbose, contenthandler = ligolw.LIGOLWContentHandler))
psd = lalseries.read_psd_xmldoc(ligolw_utils.load_filename(options.whiten_reference_psd, verbose = options.verbose, contenthandler = lalseries.PSDContentHandler))
else: <