Skip to content

Data Management and Storage

This subsection talks about the various storage spaces on Neocortex and how to use them.

Common Storage Spaces

$HOME

/jet/home/PSC_USERNAME

Warning

This storage space is subject to the Bridges-2 Storage Inode quota.

  • Here, PSC_USERNAME is your PSC username.
  • This is your personal storage space and has a cap of 25 GB. This should be used for personal files only, such as system configurations or maybe even private files like experimental Conda environments.

Danger

Please have in mind that maxing out your $HOME quota will likely lock you out of the system (storing 25GB+ under $HOME), as you might not be able to login anymore.

$PROJECT

Warning

This storage space is subject to the Bridges-2 Storage Inode quota.

If you have access to multiple projects on the PSC infrastructure, please make sure to execute the newgrp command before using the $PROJECT variable, as described in the Accounts section.

/ocean/projects/GRANT_ID

  • This storage space for storing large amounts of data and bigger files (>25GB) that are not required by the other team members. This space is the main userspace (private to each user) that charges against the group quota instead of the limited personal quota.
  • $PROJECT/../shared
    • This is a space inside the project that is writable by everyone in your specific group. If you cannot write to this folder, please ask the folder owner (the project PI) to run the following command so team members can write into that location: chmod g+sw $PROJECT/../shared.
    • This location is created by default, and it is the same type of space that was available on Phase 0 with the P## folders (also referred to as team_folder).

Neocortex-specific storage spaces

Note

The $HOME and $PROJECTS spaces are persistent, but the /local{1,2,3,4} spaces (/local1/, /local2/, /local3/, /local4/) are more ephemeral.

  • /local{1,2,3,4}
    • This storage is visible only from the SDFlex compute nodes. Using this space benefits from an extremely fast data transfer speed.
    • There are four local folders on each SDFlex node: /local1, /local2/, /local3, /local4 on each of the two SDF socket partitions. The path to these locations can also be accessed via the $LOCAL environment variable.
    • The idea of the $LOCAL space is to locally “cache” specific contents of the Ocean shared file system so the tasks performed on the SDF hosts use as many local resources as possible for the best performance.
    • Since each job task started via SLURM will have different NUMA -to-storage affinity based on the CPU cores allocated, it could be that using /localX instead of /localY for your specific task could have better performance. Having this in mind, a dynamic affinity mapping script is automatically started at the beginning of each job for setting the $LOCAL environment variable to point to the /local{1..4} location that is expected to be fastest every time.
      • Example: echo $LOCAL = /local3
    • These storage spaces are referred to as /local{1..4}, /local[1..4], or /local{1,2,3,4}; and /localX or $LOCAL. The difference between them is that when using /local and a number from 1 through 4, all of the 4 different disk spaces are being referred to; and when using /localX or $LOCAL, only one of the spaces is being addressed.

Use the following code before initializing the config with CSRunConfig to add the LOCAL environment variable:

if 'LOCAL' in os.environ:
    params['train_input']['data_dir'] = os.path.join(os.environ.get('LOCAL'), params['train_input']['data_dir'])
    params['eval_input']['data_dir'] = os.path.join(os.environ.get('LOCAL'), params['eval_input']['data_dir'])

# check_env(runconfig_params)    
# config = CSRunConfig(
#   
  • It is recommended that you move your dataset from Bridges-2 (ocean or jet) to one of the locals (/local[1-4]) on the SDF node to avoid data I/O bottleneck from Bridges-2. The mcp command should be run from inside the SDF as anywhere else it will not have access to the /local disks. In order to do that, you would have to ssh into Neocortex and start an interact command like this:

    ssh PSC_USERNAME@neocortex.psc.edu
    srun --pty bash -i
    /ocean/neocortex/tests/tools/mcp -r --direct-read --direct-write --print-stats --threads=120 /path/to/dataset/on/ocean/or/jet /local[1-4]/GRANT_ID
    
    Exit the interactive session when the data transfer is finished.

    You can use any of the /locals for this step. An important note is to transfer your dataset to both SDFs (sdf-1 and sdf-2) as when you run a slurm job to train your model, it will pick up any of the available partitions (sdf-1 or sdf-2)

Reference resources:

  • /ocean/neocortex/cerebras/
    • This space is the location in the shared-filesystem in which the different reference resources will be placed. It contains the “modelzoo” directory from Cerebras, the datasets required for running those models, the Cerebras container, example scripts, and other helpful resources.
    • This will be the primary location of these resources.
  • $LOCAL/cerebras
    • Since the main reference resources have to be accessed over the network, local copies were placed in the local storage available inside the SDF hosts. You can access the contents using this “$LOCAL/cerebras” location, as it will access the local folder that is expected to provide the highest performance for your job.
  • Cerebras container
    • This is a Singularity container created by Cerebras that contains all of the different resources required for running your code on the CS-2 boxes. The container is updated regularly and the latest version is always available in the reference folder as /ocean/neocortex/cerebras/cbcore_latest.sif
    • Please see the “Software Stack” section for more details.
  • Cerebras Datasets
    • There are some common datasets available that are used by the sample models in the Cerebras modelzoo. To find those datasets, please refer to the locations /ocean/neocortex/cerebras/ and $LOCAL/cerebras/data
  • Community datasets

    • Additionally to the datasets provided by Cerebras, there are other datasets that are made available via Ocean from the Bridges-2 supercomputer. You can find: COCO, genomics, ImageNet, MNIST, nltk, prevent_a, and other under /ocean/datasets/community
  • Where to keep your custom datasets?

    • It is advisable to store datasets under the $PROJECT/../shared folder in the Ocean shared file system; and then to copy those that are required to the /local{1,2,3,4} spaces in the SDFlex when training on the CS-2. Please remember you can access the /local/{1..4} spaces via the $LOCAL/ variable.
  • For more details regarding how to run various jobs on the system, refer to the Running jobs section of the user-guide.

Transferring Data

  • How to move datasets into the Neocortex system and back?

    • You can use the Neocortex Data Transfer Nodes (DTN) for transferring data to and from the Neocortex system. You can use rsync (preferred), scp (not recommended), or sftp commands for the same.
    • Every grant folder /ocean/projects/GRANT_ID/ has a shared file space (for everyone in the group to use) and personal user space. To use the shared space, for storing datasets and files you want to share with your team, look for the shared folder under your group folder ($PROJECT/../shared/).

    There are three ways to transfer files to/from Neocortex: rsync, scp and sftp. Below are some sample commands you can refer to.

    rsync

    rsync -PaL --chmod u+w /local-path/to/dataset PSC_USERNAME@data.neocortex.psc.edu:/ocean/projects/GRANT_ID/shared/

    where GRANT_ID is your grant id and PSC_USERNAME is your PSC username.

    Note that by default, rsync will not copy older files with the same name in place of newer files in the target directory.

    sftp file transfer is also supported.

    For more details, please visit https://www.psc.edu/resources/bridges-2/user-guide/#transferring-files