Skip to main content

Onboarding your Bunker Playbook (OB 1.1)

Download OpenAPI specification:Download

The "Onboarding your Bunker" Playbook provides the essential endpoints of the InfoSum API to securely manage data throughout the onboarding process. It allows you to import, normalize, prepare, and publish data for analysis, moving from cloud vaults to datasets while maintaining strong privacy protections.

InfoSum Support | Terms of Service | Official Documentation

About Bunkers: A Bunker is a secure private environment that stores a dataset. Only you control access to your Bunker. Collaborators can be granted permission to query your Bunker, but they will only ever see anonymized, aggregated results, never raw data. Each Bunker is hosted on its own private virtual server, encrypted and region-locked to its Cloud Vault. You may publish several datasets to different Bunkers and refresh or replace them as needed.

Tip! We recommend completing your initial onboarding in the UI first. If you have not already, please contact your InfoSum representative to arrange training. Once the setup is complete, you will be able to replicate the same steps using the API.

Task List

  1. Identify your server via an import connector configuration or ICC

  2. Import the data to your Cloud Vault using an Importer

  3. Confirm import was created successfully

  4. Create a recordset of the data’s format

  5. Confirm recordset was created successfully

  6. List Normalization Configurations

  7. Normalize the data (standardization)

  8. Confirm normalization was created successfully

  9. List Dataset Prepare Configurations

  10. Prepare the data

  11. Confirm bunker was prepared successfully

  12. Publish the data to a Bunker

  13. Confirm bunker was published successfully

Parameter Reference Table

Step Number Parameter Name Parameter Structure Where to find it When to collect
1, 4, 7, 10 Cloud Vault ID CLV.AWS_Region_selected.xx in URL of Cloud Vaults page Have ready before hand
2 Import ID IMP.xx in URL of Importing page, when an Importer has been selected Have ready before hand
3 Import Execution ID IMT.xx Response from step 2 Can be grabbed from output of step 2
4 Recordset Config ID RSC.xx Go Recordset Configs on the Cloud Vault in URL of Recordset Configs, when a config has been selected Have ready before hand
5 Recordset Execution ID MRC.xx Response from step 4 Can be grabbed from output of step 4
7 Recordset ID RSC.xx Response from step 5 Can be grabbed from output of step 5
6 Normalization Config Name text Dropdown Normalization Config on the Normalize page Have ready before hand
7 Normalization Config ID NMC.xx Response from step 5 Can be grabbed from output of step 6
8 Normalization Execution ID NMZ.xx Response from step 7 Can be grabbed from output of step 7
9 Dataset Prepare Config Name text-DD MMM YYYY - HH:MM Automations page When creating a dataset config, save the name Have ready before hand
10 Normalized Recordset ID RCS.xx Response from step 8 Can be grabbed from output of step 8
10 Dataset Prepare ID AWS_Region_selected.xx in URL of Datasets page, when a dataset is selected Have ready before hand
10 Dataset Prepare Config ID DPC.xx Response from step 9 Can be grabbed from output of step 9
11, 12, 13 Prepare/Publish Execution ID PUB.xx Response from step 10 Can be grabbed from output of step 10

Retention: Cloud Vaults serve as secure staging areas for your data and are subject to a 30-day retention policy for unpublished files. All files are encrypted at rest in AWS and must remain in the same region as their associated Bunkers.

Best Practices

  • Use Cloud Vault imports rather than local file uploads where possible for stronger privacy and easier management.
    • If you wish to use local file imports via API please reach out to your InfoSum representative to discuss your setup.
  • Assign the correct user rights, including all “Bunker Operations” rights, before onboarding data.
  • Normalize using the Global Schema to ensure consistent joins across collaborations.
  • Automate imports only after completing a successful manual onboarding.

Onboarding

1. Get a Cloud Vault.

Get the details of a specific Cloud Vault.

Authorizations:
API_Key
path Parameters
Cloud Vault ID
required
string
Example: CLV.xx

ID of Cloud Vault you would like to upload file to

header Parameters
Accept
string
Example: application/json

Responses

Response samples

Content type
application/json
{
  • "cloud_vault": {
    }
}

2. Execute Import

Executes an import for the specified import ID.

Authorizations:
API_Key
path Parameters
Import ID
required
string
Example: IMP.xx

ID of the Importer used to import data from server

header Parameters
Accept
string
Example: application/json

Responses

Response samples

Content type
application/json
{
  • "execution_id": "<string>"
}

3. Gets Import Execution

Gets a specific import execution.

Authorizations:
API_Key
path Parameters
execution_id
required
string
Example: IMT.xx

The ID of the Import execution task to retrieve.

header Parameters
Accept
string
Example: application/json

Responses

Response samples

Content type
application/json
{
  • "execution_details": {
    }
}

4. Create a Recordset.

Create a recordset from files on a cloud vault. Valid characters for recordset names: alphanumeric, '_', '-', ' '.

Authorizations:
API_Key
header Parameters
Content-Type
string
Example: application/json
Accept
string
Example: application/json
Request Body schema: application/json
object

Responses

Request samples

Content type
application/json
{
  • "cloud_vault_id": "Cloud Vault ID (CLV)",
  • "recordset_name": "RecordSet Name",
  • "config_id": "Recordset Config ID (RSC)",
  • "filepaths": [
    ]
}

Response samples

Content type
application/json
{
  • "execution_id": "<string>"
}

5. Get a Recordset Execution.

Get the details of a specific recordset execution.

Authorizations:
API_Key
path Parameters
execution_id
required
string
Example: MRC.xx

The ID of the normalization execution to retrieve.

header Parameters
Accept
string
Example: application/json

Responses

Response samples

Content type
application/json
{
  • "execution_details": {
    }
}

6. List company Normalization Configurations.

List company normalization configurations.

Authorizations:
API_Key
query Parameters
filter
string
Example: filter=name eq <Normalization Config Name>

Filter is a string containing logic that can specify a restricted set of Cloud Vaults to be returned. e.g. name co 'Test' or name co 'Acme'

The following values are permitted: name created_date_time (RFC3339 format) updated_date_time (RFC3339 format)

header Parameters
Accept
string
Example: application/json

Responses

Response samples

Content type
application/json
{
  • "configs": [
    ],
  • "pagination": {
    }
}

7. Start a Normalization.

Start a normalization

Authorizations:
API_Key
header Parameters
Content-Type
string
Example: application/json
Accept
string
Example: application/json
Request Body schema: application/json
object

Responses

Request samples

Content type
application/json
{
  • "cloud_vault_id": "Cloud Vault ID (CLV)",
  • "input_recordset_id": "RecordSet ID (RSC 2)",
  • "normalization_config_id": "Normalization Config ID (NMC)",
  • "output_recordset_name": "RecordSet Name",
  • "processing_speed": 1
}

Response samples

Content type
application/json
{
  • "execution_id": "<string>"
}

8. Get a Normalization.

Get a normalization

Authorizations:
API_Key
path Parameters
execution_id
required
string
Example: NMZ.xx

Normalization Execution ID (NMZ) is the ID of the normalization execution.

header Parameters
Accept
string
Example: application/json

Responses

Response samples

Content type
application/json
{
  • "normalization_details": {
    }
}

9. List company Dataset Prepare Configurations.

List company dataset prepare configurations

Authorizations:
API_Key
query Parameters
filter
string
Example: filter=name eq 'Dataset Prepare Config Name'

Filter is a string containing logic that can specify a restricted set of Cloud Vaults to be returned. e.g. name co 'Test' or name co 'Acme'

The following values are permitted: name created_date_time (RFC3339 format) updated_date_time (RFC3339 format)

header Parameters
Accept
string
Example: application/json

Responses

Response samples

Content type
application/json
{
  • "configs": [
    ],
  • "pagination": {
    }
}

10. Start a Prepare on normalized data.

Start a Prepare Operation

Authorizations:
API_Key
header Parameters
Content-Type
string
Example: application/json
Accept
string
Example: application/json
Request Body schema: application/json
object

Responses

Request samples

Content type
application/json
{
  • "cloud_vault_id": "Cloud Vault ID (CLV)",
  • "config_id": "Dataset Prepare Config ID (DPC)",
  • "dataset_id": "Dataset Prepare ID",
  • "recordset_id": "Normalized Recordset ID (RCS)"
}

Response samples

Content type
application/json
{
  • "execution_id": "<string>"
}

11. Get details on a Prepare operation.

Get details on a Prepare operation.

Authorizations:
API_Key
path Parameters
execution_id
required
string
Example: PUB.xx

(Required) ExecutionID is the ID of the prepare execution to fetch.

header Parameters
Accept
string
Example: application/json

Responses

Response samples

Content type
application/json
{
  • "prepare_details": {
    }
}

12. Publishes prepared data to a dataset.

Publishes prepared data to a dataset.

Authorizations:
API_Key
path Parameters
execution_id
required
string
Example: PUB.xx

ExecutionID is the id of the prepare execution to publish to the dataset.

header Parameters
Accept
string
Example: application/json

Responses

Response samples

Content type
application/json
{
  • "execution_id": "<string>"
}

13. Get details on a Publish operation.

Get details on a Publish operation.

Authorizations:
API_Key
path Parameters
execution_id
required
string
Example: PUB.xx

(Required) ExecutionID is the ID of the publish execution to fetch.

header Parameters
Accept
string
Example: application/json

Responses

Response samples

Content type
application/json
{
  • "prepare_details": {
    }
}