DeepLynx is a open-source data warehouse focused on enabling complex projects to embrace digital engineering. It accomplishes bringing digital thread and digital twins to these projects with integrations to a large collection of software systems across a project's lifecycle.
Data is stored in a graph-like format following a user-defined domain ontology. Using the provided GraphQL interface, users and applications can request exactly the data they need by using client side defined queries. This aids finding relationships between complex datasets enabling data science efforts and AI/ML.
The construction of megaprojects has consistently demonstrated challenges for project managers in regard to meeting cost, schedule, and performance requirements.
Currently, engineering teams operate in siloed tools and disparate teams. Data connections across design, procurement, construction, and operations systems are translated manually or over brittle point-to-point integrations.
This uncoordinated and disjoint data exchange across these siloes increases the risk of silent errors. These silent errors can cascade across the effort and lead to uncontrollable risk during construction, resulting in significant delays and cost overruns.
DeepLynx is a key tool in solving this problem for megaprojects by bringing those siloed efforts into an integrated platform that operates over the course of a project's lifecycle. DeepLynx integrates to widely used enterprise scale software. The list of software integrations include tools such as Innoslate for systems engineering, IBM's DOORS for requirements management, design tools such as AutoDesk's Revit, and asset management in ABB's AssetSuite.
Leveraging this rich set of integrations allows for projects to efficiently consolidate their data into a cohesive data lake. This data lake provides the foundation for digital thread and digital twin efforts.
DeepLynx
is documented in the following ways
- Wiki
- API level documentation in the form of an OpenAPI (Swagger) collection - found in the
API Documentation
folder
The easiest way to get DeepLynx up and running is to use Docker. Docker allows for easily reproducible builds and the majority of configuration tasks and database migrations are handled for you.
- Install the latest version of Docker
- Clone the DeepLynx repository
- Navigate to the DeepLynx repository in your terminal/command line/Powershell
- Type
docker compose pull
and wait for the process to finish - Type
docker compose up --attach deep-lynx
and hit enter - To terminate hit Cntrl-C or Cntrl-D
The initial startup might take a while as the operation must first fetch the pre-built containers from the internet. If you need to change any configuration values edit the .docker-env
file included in your repository
- node.js 16.x, 17.x, 18.x, 19.x (untested)
- Typescript ^4.x.x
- npm ^6.x
- Rust ^1.. (set to default stable)
- Docker ^18.x - optional - for ease of use in development
Data Source Requirements
- Required - PostgreSQL ^12.x
- Required -
pg-crypto
Postgres extension (automatically included with Postgres > 12 and in the Docker images) - TimescaleDB Postgres Extension - needed for raw data retention and time-series data
You must follow these steps in the exact order given. Failure to do so will cause Deep Lynx to either fail to launch, or launch with problems.
-
NodeJS must be installed. You can find the download for your platform here: https://nodejs.org/en/download/ note - Newer versions of Node may be incompatible with some of the following commands. The most recent version tested that works fully is 16.13.0 - the latest LTS version.
-
Clone the DeepLynx repository.
-
Change directories with
cd ./NodeLibraries/dl-fast-load
and runcargo clean && cargo build --release --message-format=json-render-diagnostics > build-output.txt
- this preps the dl-fast-load rust module for installation -
Return to the root DeepLynx directory with
cd ../../
and runnpm upgrade && npm ci
. -
Copy and rename
.env-sample
to.env
. -
Update
.env
file. See thereadme
or comments in the file itself for details. -
To build the database using docker, follow step a. To use a dedicated PostgreSQL database, follow step b. Then continue to step 8.
-
7a) Building the database using Docker:
- Ensure Docker is installed. You can find the download here: https://www.docker.com/products/docker-desktop.
- Run
npm run docker:postgres:build
to create a docker image containing a Postgres data source. - Mac users may need to create the directory to mount to the docker container at
/private/var/lib/docker/basedata
. If this directory does not exist, please create it (you may need to usesudo
as insudo mkdir /private/var/lib/docker/basedata
). - Verify that image is properly created. See the screenshot below from Docker Desktop.
- Run
npm run docker:postgres:run
to run the created docker image (For Mac users, there is an alternative commandnpm run mac:docker:postgres:run
). - Alternatively you may use
npm run docker:timescale:run
(npm run mac:docker:timescale:run
for Mac)to run a Postgres Docker image with the TimescaleDB extension already installed - to use TimescaleDB change the.env
environment variableTIMESCALEDB_ENABLED
to betrue
-
7b) Building the database using a dedicated PostgreSQL database:
- Ensure PostgreSQL is installed. You can find the download here: https://www.postgresql.org/download/. Please see this page for the latest requirements on PostgreSQL version.
- Run pgAdmin and create a new database. The database name should match whatever value is provided in the
CORE_DB_CONNECTION_STRING
of the.env
file. The default value isdeep_lynx
. - Ensure a user has been created that also matches the
CORE_DB_CONNECTION_STRING
and that the user's password has been set appropriately. The default username ispostgres
and the default password isdeeplynxcore
.
-
Run
npm run build
to build the internal modules and bundled administration GUI. Note You must re-run this command if you make changes to the administration GUI. -
Run
npm run watch
ornpm run start
to start the application. See thereadme
for additional details and available commands. This command starts a process that only ends when a user terminates with Cntrl+C or Cntrl+D - you will see a constant feed of logs from this terminal once you have started Deep Lynx. This is normal. Changes to the source code of Deep Lynx will be captured if you run the application with thenpm run watch
command.
Note: DeepLynx ships with a Vue single page application which serves as the primary UI for the DeepLynx system. You can run this separately (and it's recommended to do so if you're developing it).
The bundled admin web GUI can be accessed at {{your base URL}}
- default is localhost:8090
DeepLynx ships with the capability to utilize a Postgres plugin called TimescaleDB. We use this for the storage of time-series data as well as a potential target for raw data retention. This is a powerful tool and you must have it enabled in order to store time-series data on nodes.
- Change the
TIMESCALEDB_ENABLED
environment variable to readtrue
- Restart the application.
Note: Once you enable TimescaleDB you cannot disable it. Please make sure you absolutely need this extension of DeepLynx before taking steps to enable.
The bundled admin web GUI can be accessed at {{your base URL}}
- default is localhost:8090
This application's configuration relies on environment variables of its host system. It is best to rely on your CI/CD pipeline to inject those variables into your runtime environment.
In order to facilitate local development, a method has been provided to configure the application as if you were setting environment variables on your local machine. Including a .env
file at the projects root and using the npm run watch
, npm run start
, or any of the npm run docker:*
commands will start the application loading the listed variables in that file. See the .env-sample
file included as part of the project for a list of required variables and formatting help.
A database migration step takes place each time you launch the application. This ensures that your local database always has the correct schema for your branch.
DeepLynx ships with the capability to utilize a Postgres plugin called TimescaleDB. We use this for the storage of time-series data as well as a potential target for raw data retention. This is a powerful tool and you must have it enabled in order to store time-series data on nodes.
- Change the
TIMESCALEDB_ENABLED
environment variable to readtrue
- Restart the application
Note: Once you enable TimescaleDB you cannot disable it. Please make sure you absolutely need this extension of DeepLynx before taking steps to enable.
This application uses Mocha and Chai for its unit and integration tests. Visit their respective websites and documentation to learn more about how to use and develop tests with their platforms.
IMPORTANT NOTE
If you decide to test graph functionality (Gremlin functionality in particular) in isolation or use something other than a CosmosDB or CosmosDB emulator you must ensure that the DATA_SOURCE_GRAPHSON_V1
environment variable is left blank. Failure to do so means you might be communicating in an unsupported format, or an unsupported combination of formats.
You must also have run Deep Lynx at least once in order to run the testing suite.
Below is a list of all npm run
commands as listed in the package.json
file.
docker:api:build
Creates a docker image of Deep Lynx injecting the .env file for configuration.docker:api:run
Runs previously created Deep Lynx image.docker:api:clean
Stops the Deep Lynx Docker container run by the command above and deletes the container and image.docker:postgres:build
Creates a Docker image containing a Postgres 12 data source.docker:postgres:run
Runs previously created Postgres image.docker:postgres:clean
Stops the Postgres Docker container run by the command above and deletes the container and image.docker:timescale:run
Runs a Postgres 12 Docker container with TimescaleDB already installed.build
Compiles the applicationstart
Runs the compiled applicationwatch
Starts the application and rebuilds it each time you make a change to the code. Note: this command will not rebuild the bundled Admin Web Applicationtest
: Runs all tests using the.env
file to configure application and tests Note: You must have run the application at least once so that the database migration took place correctly.
There is a lot more information about Deep Lynx, and it's capabilities in its Wiki. We highly recommend you start there if you have questions or need to figure out how best to utilize Deep Lynx in your project.