CMS 3D CMS Logo

MatrixUtil.py
Go to the documentation of this file.
1 class Matrix(dict):
2  def __setitem__(self,key,value):
3  if key in self:
4  print "ERROR in Matrix"
5  print "overwritting",key,"not allowed"
6  else:
7  self.update({float(key):WF(float(key),value)})
8 
9  def addOverride(self,key,override):
10  self[key].addOverride(override)
11 
12 #the class to collect all possible steps
13 class Steps(dict):
14  def __setitem__(self,key,value):
15  if key in self:
16  print "ERROR in Step"
17  print "overwritting",key,"not allowed"
18  import sys
19  sys.exit(-9)
20  else:
21  self.update({key:value})
22  # make the python file named <step>.py
23  #if not '--python' in value: self[key].update({'--python':'%s.py'%(key,)})
24 
25  def overwrite(self,keypair):
26  value=self[keypair[1]]
27  self.update({keypair[0]:value})
28 
29 class WF(list):
30  def __init__(self,n,l):
31  self.extend(l)
32  self.num=n
33  #the actual steps of this WF
34  self.steps=[]
35  self.overrides={}
36  def addOverride(self,overrides):
37  self.overrides=overrides
38 
39  def interpret(self,stepsDict):
40  for s in self:
41  print 'steps',s,stepsDict[s]
42  steps.append(stepsDict[s])
43 
44 
45 
46 def expandLsInterval(lumis):
47  return range(lumis[0],(lumis[1]+1))
48 
50 jsonFile2015 = findFileInPath("DPGAnalysis/Skims/data/Cert_13TeV_16Dec2015ReReco_Collisions15_25ns_50ns_JSON.txt")
51 jsonFile2016 = findFileInPath("DPGAnalysis/Skims/data/Cert_271036-274240_13TeV_PromptReco_Collisions16_JSON.txt")
52 
53 import json
54 with open(jsonFile2015) as data_file:
55  data_json2015 = json.load(data_file)
56 
57 with open(jsonFile2016) as data_file:
58  data_json2016 = json.load(data_file)
59 
60 # return a portion of the 2015 golden json
61 # LS for a full run by default; otherwise a subset of which you determined the size
62 def selectedLS(list_runs=[],maxNum=-1,l_json=data_json2015):
63  # print "maxNum is %s"%(maxNum)
64  if type(list_runs[0]) !=int:
65  print "ERROR: list_runs must be a list of intergers"
66  return None
67  local_dict = {}
68  ls_count = 0
69 
70  for run in list_runs:
71  if str(run) in l_json.keys():
72  # print "run %s is there"%(run)
73  runNumber = run
74  # print "Doing lumi-section selection for run %s: "%(run)
75  for LSsegment in l_json[str(run)] :
76  # print LSsegment
77  ls_count += (LSsegment[-1] - LSsegment[0] + 1)
78  if (ls_count > maxNum) & (maxNum != -1):
79  break
80  # return local_dict
81  if runNumber in local_dict.keys():
82  local_dict[runNumber].append(LSsegment)
83  else:
84  local_dict[runNumber] = [LSsegment]
85  # print "total LS so far %s - grow %s"%(ls_count,local_dict)
86  #local_dict[runNumber] = [1,2,3]
87  else:
88  print "run %s is NOT present in json %s\n\n"%(run, l_json)
89  # print "++ %s"%(local_dict)
90 
91  if ( len(local_dict.keys()) > 0 ) :
92  return local_dict
93  else :
94  print "No luminosity section interval passed the json and your selection; returning None"
95  return None
96 
97 # print "\n\n\n THIS IS WHAT I RETURN: %s \n\n"%( selectedLS([251244,251251]) )
98 
99 
100 
101 
102 InputInfoNDefault=2000000
104  def __init__(self,dataSet,label='',run=[],ls={},files=1000,events=InputInfoNDefault,split=10,location='CAF',ib_blacklist=None,ib_block=None) :
105  self.run = run
106  self.ls = ls
107  self.files = files
108  self.events = events
109  self.location = location
110  self.label = label
111  self.dataSet = dataSet
112  self.split = split
113  self.ib_blacklist = ib_blacklist
114  self.ib_block = ib_block
115 
116  def das(self, das_options):
117  if len(self.run) is not 0 or self.ls:
118  queries = self.queries()[:3]
119  if len(self.run) != 0:
120  command = ";".join(["dasgoclient %s --query '%s'" % (das_options, query) for query in queries])
121  else:
122  lumis = self.lumis()
123  commands = []
124  while queries:
125  commands.append("dasgoclient %s --query 'lumi,%s' --format json | das-selected-lumis.py %s " % (das_options, queries.pop(), lumis.pop()))
126  command = ";".join(commands)
127  command = "({0})".format(command)
128  else:
129  command = "dasgoclient %s --query '%s'" % (das_options, self.queries()[0])
130 
131  # Run filter on DAS output
132  if self.ib_blacklist:
133  command += " | grep -E -v "
134  command += " ".join(["-e '{0}'".format(pattern) for pattern in self.ib_blacklist])
135  command += " | sort -u"
136  return command
137 
138  def lumiRanges(self):
139  if len(self.run) != 0:
140  return "echo '{\n"+",".join(('"%d":[[1,268435455]]\n'%(x,) for x in self.run))+"}'"
141  if self.ls :
142  return "echo '{\n"+",".join(('"%d" : %s\n'%( int(x),self.ls[x]) for x in self.ls.keys()))+"}'"
143  return None
144 
145  def lumis(self):
146  query_lumis = []
147  if self.ls:
148  for run in self.ls.keys():
149  run_lumis = []
150  for rng in self.ls[run]: run_lumis.append(str(rng[0])+","+str(rng[1]))
151  query_lumis.append(":".join(run_lumis))
152  return query_lumis
153 
154  def queries(self):
155  query_by = "block" if self.ib_block else "dataset"
156  query_source = "{0}#{1}".format(self.dataSet, self.ib_block) if self.ib_block else self.dataSet
157 
158  if self.ls :
159  the_queries = []
160  #for query_run in self.ls.keys():
161  # print "run is %s"%(query_run)
162  # if you have a LS list specified, still query das for the full run (multiple ls queries take forever)
163  # and use step1_lumiRanges.log to run only on LS which respect your selection
164 
165  # DO WE WANT T2_CERN ?
166  return ["file {0}={1} run={2}".format(query_by, query_source, query_run) for query_run in self.ls.keys()]
167  #return ["file {0}={1} run={2} site=T2_CH_CERN".format(query_by, query_source, query_run) for query_run in self.ls.keys()]
168 
169 
170  #
171  #for a_range in self.ls[query_run]:
172  # # print "a_range is %s"%(a_range)
173  # the_queries += ["file {0}={1} run={2} lumi={3} ".format(query_by, query_source, query_run, query_ls) for query_ls in expandLsInterval(a_range) ]
174  #print the_queries
175  return the_queries
176 
177  if len(self.run) is not 0:
178  return ["file {0}={1} run={2} site=T2_CH_CERN".format(query_by, query_source, query_run) for query_run in self.run]
179  #return ["file {0}={1} run={2} ".format(query_by, query_source, query_run) for query_run in self.run]
180  else:
181  return ["file {0}={1} site=T2_CH_CERN".format(query_by, query_source)]
182  #return ["file {0}={1} ".format(query_by, query_source)]
183 
184  def __str__(self):
185  if self.ib_block:
186  return "input from: {0} with run {1}#{2}".format(self.dataSet, self.ib_block, self.run)
187  return "input from: {0} with run {1}".format(self.dataSet, self.run)
188 
189 
190 # merge dictionaries, with prioty on the [0] index
191 def merge(dictlist,TELL=False):
192  import copy
193  last=len(dictlist)-1
194  if TELL: print last,dictlist
195  if last==0:
196  # ONLY ONE ITEM LEFT
197  return copy.copy(dictlist[0])
198  else:
199  reducedlist=dictlist[0:max(0,last-1)]
200  if TELL: print reducedlist
201  # make a copy of the last item
202  d=copy.copy(dictlist[last])
203  # update with the last but one item
204  d.update(dictlist[last-1])
205  # and recursively do the rest
206  reducedlist.append(d)
207  return merge(reducedlist,TELL)
208 
209 def remove(d,key,TELL=False):
210  import copy
211  e = copy.deepcopy(d)
212  if TELL: print "original dict, BEF: %s"%d
213  del e[key]
214  if TELL: print "copy-removed dict, AFT: %s"%e
215  return e
216 
217 
218 #### Standard release validation samples ####
219 
220 stCond={'--conditions':'auto:run1_mc'}
221 def Kby(N,s):
222  return {'--relval':'%s000,%s'%(N,s)}
223 def Mby(N,s):
224  return {'--relval':'%s000000,%s'%(N,s)}
225 
226 def changeRefRelease(steps,listOfPairs):
227  for s in steps:
228  if ('INPUT' in steps[s]):
229  oldD=steps[s]['INPUT'].dataSet
230  for (ref,newRef) in listOfPairs:
231  if ref in oldD:
232  steps[s]['INPUT'].dataSet=oldD.replace(ref,newRef)
233  if '--pileup_input' in steps[s]:
234  for (ref,newRef) in listOfPairs:
235  if ref in steps[s]['--pileup_input']:
236  steps[s]['--pileup_input']=steps[s]['--pileup_input'].replace(ref,newRef)
237 
238 def addForAll(steps,d):
239  for s in steps:
240  steps[s].update(d)
241 
242 
243 def genvalid(fragment,d,suffix='all',fi='',dataSet=''):
244  import copy
245  c=copy.copy(d)
246  if suffix:
247  c['-s']=c['-s'].replace('genvalid','genvalid_'+suffix)
248  if fi:
249  c['--filein']='lhe:%d'%(fi,)
250  if dataSet:
251  c['--filein']='das:%s'%(dataSet,)
252  c['cfg']=fragment
253  return c
254 
255 
Definition: merge.py:1
def interpret(self, stepsDict)
Definition: MatrixUtil.py:39
def __setitem__(self, key, value)
Definition: MatrixUtil.py:14
def genvalid(fragment, d, suffix='all', fi='', dataSet='')
Definition: MatrixUtil.py:243
def overwrite(self, keypair)
Definition: MatrixUtil.py:25
def replace(string, replacements)
def __init__(self, n, l)
Definition: MatrixUtil.py:30
def lumiRanges(self)
Definition: MatrixUtil.py:138
def findFileInPath(theFile)
def expandLsInterval(lumis)
Definition: MatrixUtil.py:46
def addForAll(steps, d)
Definition: MatrixUtil.py:238
def __setitem__(self, key, value)
Definition: MatrixUtil.py:2
def das(self, das_options)
Definition: MatrixUtil.py:116
static std::string join(char **cmd)
Definition: RemoteFile.cc:18
def selectedLS(list_runs=[], maxNum=-1, l_json=data_json2015)
Definition: MatrixUtil.py:62
def remove(d, key, TELL=False)
Definition: MatrixUtil.py:209
def __init__(self, dataSet, label='', run=[], ls={}, files=1000, events=InputInfoNDefault, split=10, location='CAF', ib_blacklist=None, ib_block=None)
Definition: MatrixUtil.py:104
def changeRefRelease(steps, listOfPairs)
Definition: MatrixUtil.py:226
#define update(a, b)
def Mby(N, s)
Definition: MatrixUtil.py:223
def addOverride(self, key, override)
Definition: MatrixUtil.py:9
def Kby(N, s)
Standard release validation samples ####.
Definition: MatrixUtil.py:221
def addOverride(self, overrides)
Definition: MatrixUtil.py:36
How EventSelector::AcceptEvent() decides whether to accept an event for output otherwise it is excluding the probing of A single or multiple positive and the trigger will pass if any such matching triggers are PASS or EXCEPTION[A criterion thatmatches no triggers at all is detected and causes a throw.] A single negative with an expectation of appropriate bit checking in the decision and the trigger will pass if any such matching triggers are FAIL or EXCEPTION A wildcarded negative criterion that matches more than one trigger in the trigger list("!*","!HLTx*"if it matches 2 triggers or more) will accept the event if all the matching triggers are FAIL.It will reject the event if any of the triggers are PASS or EXCEPTION(this matches the behavior of"!*"before the partial wildcard feature was incorporated).Triggers which are in the READY state are completely ignored.(READY should never be returned since the trigger paths have been run
def merge(dictlist, TELL=False)
Definition: MatrixUtil.py:191