Advanced Modelling Use Cases

Every database is different, and sometimes one or more of the following edge case scenarios may need to be considered when modelling the data.

Multiple Datasets Mapped to the Same Table

This is typically needed when one table represents multiple logical entities, and each entity should be represented by a separate dataset in the LDM.

While the LDM Modeler supports mapping of multiple datasets to the same table, publishing an LDM with such mapping fails.

To avoid this issue, create multiple views on top of the table and map each dataset to a separate view.

For example, you have two tables, users and tickets.

  • The users table contains ticket creators and assignees.
  • The tickets table contains the assignee_id and creator_id columns.

To avoid mapping multiple datasets to the users table, do the following:

  1. In the database, create two views on top of the users table: v_users_assignees and v_users_creators.
  2. In the LDM, create three datasets: assignees, creators, and tickets.
  3. Map the tickets dataset to the tickets table.
  4. Map the assignees dataset to the v_users_assignees view.
  5. Map the creators dataset to the v_users_creators view.
  6. Create a relationship from the assignees dataset to the the tickets dataset using the assignee_id column as a primary key in the assignees dataset.
  7. Create a relationship from the creators dataset to the tickets dataset using the creator_id column as a primary key in the creators dataset.

No Single-column Primary Key in Tables

While the LDM Modeler supports setting multi-column primary keys in datasets, publishing an LDM with multi-column primary keys fails.

To avoid this issue, create one-attribute primary keys in your database:

  1. Concatenate the table columns that comprise the primary key in the table.
  2. Calculate a hash value for the concatenated columns.
  3. Use the hash value as a one-attribute primary key.

Multiple PDMs Generated from the Same Data Source

If you need to prepare different PDMs from the same data source to serve various business needs, we recommend that you create multiple data sources from the same database and generate a PDM for each data source separately.

To narrow down the scope of the database to scan, you can create views with different prefixes and only scan the database with a specific prefix for each data source. Also, you can define different credentials/privileges or schema for each data source.

Create a Physical Data Model Manually

To generate the PDM using API, do the following:

  1. Scan the relational model of your database.
  2. Store the PDM under the corresponding data source.

Scan the Relational Model of Your Database

To scan the relational model in your database, submit a POST request to api/v1/actions/dataSources/<data-source-id>/scan. <data-source-id> is the ID of the data source that corresponds to the database you want to scan.

By default, all the tables, views, and columns in your database are scanned, and a JSON file with a declarative definition of the PDM is generated. To generate a more accurate PDM, especially when your database contains a large amount of tables/views, you can do the following:

  • Change tables, views, and columns so that the database contains only the entities that have to be included in the PDM.

  • Fine-tune the API request to narrow down the scope of the database objects to scan:

    • Define what to scan: tables, views, or both.
    • Define the prefix to scan only the tables/views whose names contain this prefix.

You can also prepare the tables/views for use in complex analytical scenarios, such as the one described in this article.

Example:

The PDM definition will be saved to the pdm.json file.

Bash
PowerShell 7
curl $HOST_URL/api/v1/actions/dataSources/<data-source-id>/scan \
  -H "Content-Type: application/json" \
  -H "Accept: application/json" \
  -H "Authorization: Bearer $API_TOKEN" \
  -X POST \
  -d '{"separator": "__", "scanTables": false, "scanViews": true}' \
  | jq . > pdm.json
Invoke-RestMethod -Method Post -Uri "$HOST_URL/api/v1/actions/dataSources/<data-source-id>/scan" `
  -ContentType 'application/json' `
  -H @{ 
    'Accept' = 'application/json' 
    'Authorization' = 'Bearer $API_TOKEN' 
  } `
  -Body '{"separator": "__", "scanTables": false, "scanViews": true}' | ConvertTo-Json > pdm.json

Store the PDM under the Corresponding Data Source

To store the generated PDM under the corresponding data source, submit a PUT request to /api/v1/layout/dataSources/<data-source-id>/physicalModel. <data-source-id> is the ID of the data source that corresponds to the scanned database. In the API request, specify the JSON file with the PDM definition.

Example:

Bash
PowerShell 7
curl $HOST_URL/api/v1/layout/dataSources/<data-source-id>/physicalModel \
  -H "Authorization: Bearer $API_TOKEN" \
  -H "Content-Type: application/json" \
  -X PUT -d @pdm.json
Invoke-RestMethod -Method Put -Uri "$HOST_URL/api/v1/layout/DataSources/<data-source-id>/physicalModel" `
  -ContentType 'application/json' `
  -H @{ 'Authorization' = 'Bearer $API_TOKEN' } `
  -InFile pdm.json

Once you have stored the PDM, build the LDM.

  • If your database is prepared for building analytics and does not contain the complex analytical scenarios, generate the LDM automatically from the stored PDM.
  • Otherwise, create the LDM manually in the LDM Modeler.