Skip to content

Frequently Asked Questions

Getting Started

Who is eligible to request an OSG account?
Any researcher affiliated with a U.S. institution (college, university, national laboratory or research foundation) is eligible to use OSG resources for their work. Researchers outside of the U.S. with affiliations to U.S. groups may be eligible for membership if they are sponsored by a collaborator within the U.S.
How do I request an OSG account?
Please visit our website for the most up-to-date information on requesting an account. Once your account request has been received, a Research Computing Facilitator will contact you within one business day to arrange a meeting to learn about your computational goals and to create your account.
How do I change the project my jobs are affiliated with?
The OSG team assigns individual user accounts to "projects" upon account creation. These projects are a way to track usage hours and capture information about the types of research running on OSG resources. A project typically corresponds to a research group headed by a single PI, but can sometimes represent a long-term multi-institutional project or some other grouping. If you only belong to a single project, that project will be charged automatically when you submit jobs. Run the following command to see a list of projects you belong to:

$ grep $USER /etc/condor/UserToProjectMap.txt

If need to run jobs under a different project you are a member of, you can manually set the project for those jobs by putting this option in the submit file:

+ProjectName="ProjectName"
Can I use my ACCESS allocation?
There are two ways OSG interfaces with ACCESS:
  1. You can get an allocation for the OSPool. This will allow you to run OSPool jobs and have the usage charged to your ACCESS credits, and can be useful if you already have an allocation. If you only need to use OSG resources, we recommend you come directly to our system.
  2. You can manage your workloads on the OSPool access points, and run those jobs on other ACCESS resources. This is a capability still in development.

Workshops and Training

Do you offer training sessions and workshops?
We offer virtual trainings twice-a-month, as well as an annual, week-long summer school for OSG users. We also participate in additional external conferences and events throughout the year. Information about upcoming and past events, including workshop dates and locations, is available on our website.
Who may attend OSG workshops?
Workshops are available to any researcher affiliated with a U.S. academic, non-profit, or government institution.
How to cite or acknowledge OSG?
Whenever you make use of OSG resources, services or tools, we request you acknowledge OSG in your presentations and publications using the informtion provided on the Acknowledging the OSG Consortium page.

Software

What software packages are available?
In general, we support most software that fits the distributed high throughput computing model (e.g., open source). Users are encouraged to download and install their own software on our Access Points.

Additionally, users may install their software into a Docker container which can run on OSG as an Apptainer image or use one of our existing containers. See the Software guides on the OSPool documentation website for more information.
Are there any restrictions on installing commercial softwares?
We can only *directly* support software that is freely distributable. At present, we do not have or support most commercial software due to licensing issues. (One exception is running MATLAB standalone executables which have been compiled with the MATLAB Compiler Runtime). Software that is licensed to individual users (and not to be shared between users) can be staged within the user's `/home` or `/protected` directories, but should not be staged in OSG's `/public` data staging locations. See OSPool policies for more information. Please get in touch with any questions about licensed software.
Can I request for system wide installation of the open source software useful for my research?
We recommend users use Docker or Apptainer containers if jobs require system wide installations of software. Visit the OSPool Documentation website to learn more about creating your own container.

Running Jobs

What type of computation is a good match or NOT a good match for the OSG's Open Science Pool?
The OSG provides computing resources through the Open Science Pool for high throughput computing workloads. You can get the most of out OSG resources by breaking up a single large computational task into many smaller tasks for the fastest overall turnaround. This approach can be invaluable in accelerating your computational work and thus your research. Please see our Computation on the Open Science Pool page for more details on how to determine if your work matches up well with OSG's high throughput computing model.
What job scheduler is being used on the Open Science Pool?
We use a task scheduling software called HTCondor to schedule and run jobs.
How do I submit a computing job?
Jobs are submitted via HTCondor scheduler. Please see our Roadmap to HTC Workload Submission guide for more details on submitting and managing jobs.
How many jobs can I have in the queue?
The number of jobs that are submitted to the queue by any one user cannot not exceed 10,000 without adding a special statement to the submit file. If you have more jobs than that, we ask that you include the following statement in your submit file:

max_idle = 2000

This is the maximum number of jobs that you will have in the "Idle" or "Held" state for the submitted batch of jobs at any given time. Using a value of 2000 will ensure that your jobs continue to apply a constant pressure on the queue, but will not fill up the queue unnecessarily (which helps the scheduler to perform optimally).
How do I view usage metrics for my project?
The project's resource usage appears in the OSG accounting system, GRid ACcounting Collector (GRACC). Additional dashboards are available to help filter information of interest.

At the top of that dashboard, there is a set of filters that you can use to examine the number of hours used by your project, or your institution.
Why specify +JobDurationCategory in the HTCondor submit file?
To maximize the value of the capacity contributed by the different organizations to the OSPool, users are requested to identify a duration categories for their jobs. These categories should be selected based upon test jobs (run on the OSPool) and allow for more effective scheduling of the capacity contributed to the pool.

Every job submitted from an OSG-managed access point must be labeled with a Job Duration Category upon submission. By knowing the expected job duration, OSG will be able to direct longer-running jobs to resources that are faster and are interrupted less, while shorter jobs can run across more of the OSPool for better overall throughput.

Jobs with single executions longer than 20 hours in tests on the OSPool should not be submitted, without self-checkpointing.

Details on how to specify +JobDurationCategory can be found in our Overview: Submit Jobs to the OSPool using HTCondor and Roadmap to HTC Workload Submission guides.

Data Storage and Transfer

What is the best way to process large volume of data?
There may be more than one solution that is available to researchers to process large amounts of data. Contact a Facilitator at for a free, individual consulation to learn about your options.
How do I transfer my data to and from OSG Access Points?
You can transfer data using `scp`, `rsync`, or other common Unix tools. See Using scp To Transfer Files for more details.
Is there any support for private data?
Data stored in `/protected` and in `/home` is not publically accessible. Sensitive data, such as HIPAA data, is not allowed to be uploaded or analyzed using OSG resources.
Is data backed up on OSG resources?
Our data storage locations are not backed up nor are they intended for long-term storage. If the data is not being used for active computing work, it should not be stored on OSG systems.
Can I get a quota increase?
Yes. Contact [email protected] if you think you'll need a quota increase for `/home`, `/public`, or `/protected` to accommodate a set of concurrently-running jobs. We can suppport very large amounts of data, the default quotas are just a starting point.
Will I get notified about hitting quota limits?
The best place to see your quota status is in the login message.

Workflow Management

How do I run and manage complex workflows?
For workflows that have multiple steps and/or multiple files, we advise using a workflow management system. A workflow management system allows you to define different computational steps in your workflow and indicate how inputs and outputs should be transferred between these steps. Once you define a workflow, the workflow management system will then run your workflow, automatically retrying failed jobs and transferrring files between different steps.
What workflow management systems are recommended on OSG?
We support DAGMan and Pegasus for workflow management.