|
| mps_fire.action |
|
| mps_fire.args = parser.parse_args(sys.argv[1:]) |
|
| mps_fire.backupCfgPath = os.path.join(Path, mergeCfg+".bak") |
|
| mps_fire.backupScriptPath = os.path.join(Path, "theScript.sh.bak") |
|
| mps_fire.cfgPath = os.path.join(Path, mergeCfg) |
|
string | mps_fire.command = 'cat '+backupScriptPath+' | grep CONFIG_FILE | head -1 | awk -F"/" \'{print $NF}\'' |
|
string | mps_fire.curJobName = 'm'+str(nMerge)+'_'+theJobName |
|
| mps_fire.default |
|
| mps_fire.dest |
|
| mps_fire.dircontents = os.listdir(Path) |
|
| mps_fire.end |
|
bool | mps_fire.fire_htcondor = False |
|
| mps_fire.help |
|
| mps_fire.i = lib.nJobs |
|
string | mps_fire.inCfgPath = theJobData+'/'+lib.JOBDIR[0]+'/the.py' |
|
| mps_fire.int |
|
bool | mps_fire.invalid_id = False |
|
| mps_fire.job_id |
|
| mps_fire.job_mask = lib.JOBDIR |
|
def | mps_fire.job_submit_file = write_HTCondor_submit_file_mille(Path, scriptPath, lib,os.path.join(Path,".user_proxy")) |
|
int | mps_fire.jobNumFrom1 = i+1 |
|
| mps_fire.lib = mpslib.jobdatabase() |
|
| mps_fire.match = re.search(r"1 job\(s\) submitted to cluster (\d+)\.", result) |
|
| mps_fire.maxJobs |
|
| mps_fire.mergeCfg = subprocess.check_output(command, stderr=subprocess.STDOUT, shell=True) |
|
bool | mps_fire.mergeOK = True |
|
| mps_fire.nargs |
|
| mps_fire.nMerge = i-lib.nJobs |
|
int | mps_fire.nSub = 0 |
|
| mps_fire.parser |
|
| mps_fire.Path = os.path.join(theJobData,lib.JOBDIR[i]) |
|
| mps_fire.resources = lib.get_class('mille') |
|
| mps_fire.result |
|
| mps_fire.schedinfo = subprocess.check_output(["myschedd","show"]) |
|
| mps_fire.scriptPath = os.path.join(Path, "theScript.sh") |
|
string | mps_fire.submission = "condor_submit -batch-name %s %s"%\ |
|
| mps_fire.theJobData = os.path.join(os.getcwd(), "jobData") |
|
string | mps_fire.theJobName = 'mpalign' |
|
| mps_fire.type |
|
| mps_fire.weight_conf = cPickle.load(f) |
|