tool.registry.home=cipres tool.resource.cluster=TSCC tscc.login=tscc tscc.fileHost=tscc tscc.submit=submit_v2.py tscc.check=checkjobs_v2.py tscc.cancel=delete_v2.py tscc.rc=/home/cipres_test/ngbw/contrib/scripts/workbench.rc tscc.project=cipres_test tscc.workspace=/projects/ps-ngbt/backend/tscc_workspace tool.resource.host=billiards host=billiards # directory must be owned by "bwbatch" workspace=/archive/science/ngbt/portal2prod/workspace/bwbatch_jobs/ teragrid.trestles.workspace=/projects/ps-ngbt/backend/trestles_workspace teragrid.trestles.host=trestles-dm1.sdsc.edu teragrid.trestles.user= teragrid.trestles.login=cipres@trestles.sdsc.edu teragrid.trestles.rc=/projects/ps-ngbt/home/cipres/.bash_profile teragrid.trestles.submit=submit_v2.py teragrid.trestles.check=checkjobs_v2.py teragrid.trestles.cancel=delete_v2.py trestles.filehandler=org.ngbw.sdk.tool.LocalFileHandler teragrid.gordon.workspace=/projects/ps-ngbt/backend/gordon_workspace teragrid.gordon.host=trestles-dm1.sdsc.edu teragrid.gordon.user= teragrid.gordon.login=cipres@gordon.sdsc.edu teragrid.gordon.rc=/projects/ps-ngbt/home/cipres/.bash_profile teragrid.gordon.submit=gordon_submit_v2.py teragrid.gordon.check=checkjobs_v2.py teragrid.gordon.cancel=delete_v2.py gordon.filehandler=org.ngbw.sdk.tool.LocalFileHandler # Set to "true" to disable teragrid tools, "false" to enable teragrid tools. teragrid.tools.disable=false # #################################################################################### # GLOBUS Teragrid Properties # #################################################################################### # Units for the following are in minutes (1 hr=60, 1 day = 1440, 3 days = 4320, 7 days = 10080, etc) # The longest proxy we can get from myproxy with our current credential is 7 days so # there's no point trying to run jobs longer than that. Lonestar only allows 2 days anyway. grid.job.lifetime.max = 10080 # Globus process worker polls very frequently the first minute, then at quick.poll.interval for # several minutes and finally at poll.interval. grid.job.poll.interval = 60 grid.job.medium.poll.interval = 5 grid.job.quick.poll.interval = 1 # Lifetime of credential in hours. min.lifetime must be less than lifetime. # With the credential we're using (a default teragrid credential) 7 days = 168 hrs is the max # we can get. grid.myproxy.lifetime = 168 grid.myproxy.min.lifetime = 144 grid.myproxyfile=/tmp/x509_cipres_prod # #################################################################################### # END GLOBUS Teragrid Properties # #################################################################################### portal.name=Cipres Portal job.email.from=CIPRES Services job.email.enable=true emailHost=smtp.ucsd.edu # maxPoolSize needs to be at least ( 2 * max(loadResults.pool.size, submitJob.pool.size) ) + 1 # to avoid deadlock. Also make sure that in ssl.properties, maxconn for the hosts we're using # is at least 2 * max(loadResults.pool.size, submitJob.pool.size) database.maxPoolSize=25 database.minPoolSize=2 ssl.maxconn=25 # Poll remote host for completed jobs every 10 minutes checkJobs.poll.seconds=1800 checkjobs.maxDays = 30 loadResults.poll.seconds=60 loadResults.pool.size=10 loadResults.retries=40 submitJob.poll.seconds=10 submitJob.pool.size=10 submitJob.retries=3 disabled.resources.file=/archive/science/ngbt/portal2prod/disabled_resources accounting.period.start=2013-07-01 iplant.charge.number=TG-MCB110022 cipres.charge.number=TG-DEB090011 logs=/fs/portal2prod/logs portal.url=http://www.phylo.org/portal2 rest.callback.url=-K ~/.jobcurl.rc https://www.phylo.org/cipresrest/v1/admin/updateJob portal.callback.url=http://www.phylo.org/portal2/taskupdate.action server_url=https://www.phylo.org/cipresrest/v1/job rest_url=https://www.phylo.org/cipresrest/v1 database.url=jdbc:log4jdbc:mysql://mysql.sdsc.edu:3316/cipres database.fileRoot=/archive/science/ngbt/db_documents/ngbw_prod database.host=mysql.sdsc.edu database.port=3316 database.name=cipres testpass=test pycipres_admin=pycipres_admin pycipres_app=pycipres pycipres_eu=pycipres_eu guidemo_admin=guidemo_admin guidemo_app=guidemo guidemo2_admin=guidemo2_admin guidemo2_app=guidemo2 xsede_su_limit=30000 other_su_limit=30000 submitted_today_limit=50 active_limit=50 concurrent_limit=10 # no_limits=1