It is possible to upload data on Brainlife. Datasets can be public or private (it depends on the nature of the project it is linked to).

Store data on BL

There are two ways to store data on BL:

Upload data on BL

To upload data, go to the Processes page and then in the Archive tab. In the right corner, you see a green button “Upload data”. Then, select the datatype of your data and fill in the blanks (see BL documentation).

But as you can see, the meg/fifdatatype is not listed when you want to select this datatype for your dataset: you can’t upload your data using the BL GUI, you have to use the BL CLI.

No need to install all of the dependencies, you can use the BL CLI hosted as a docker container via Singularity by running the following command:

$ singularity run docker://brainlife/cli login

The result is:

For more details, see the BL documentation.

In order to upload a data object, Brainlife requires that you supply a project and datatype associated with it it:

  1. get the ID of your project

    $ singularity run docker://brainlife/cli project query --admin <admin>

  2. get the ID of the datatype of your project

    $ singularity run docker://brainlife/cli datatype query --query <datatype>

Once you get these two ID, you can upload your data. To do so, all that is required is a directory to upload, a project id, and a datatype id (but other options exist):

$ singularity run docker://brainlife/cli data upload --directory <path> --project <projectid> --datatype <datatypeid>

It can be heavy to always write singularity run docker://brainlife/cli when using BL CLI, so you can add this command to your ~/.bashrc:

function bl {
    singularity run docker://brainlife/cli $@
}

Besides, it may be possible to create a bash script that upload automatically your data (to do).