Skip to content

GitLab

  • Menu
Projects Groups Snippets
  • Help
    • Help
    • Support
    • Community forum
    • Submit feedback
  • Sign in / Register
  • E External Data Framework
  • Project information
    • Project information
    • Activity
    • Labels
    • Members
  • Repository
    • Repository
    • Files
    • Commits
    • Branches
    • Tags
    • Contributors
    • Graph
    • Compare
    • Locked Files
  • Issues 74
    • Issues 74
    • List
    • Boards
    • Service Desk
    • Milestones
    • Iterations
    • Requirements
  • Merge requests 0
    • Merge requests 0
  • CI/CD
    • CI/CD
    • Pipelines
    • Jobs
    • Schedules
    • Test Cases
  • Deployments
    • Deployments
    • Environments
    • Releases
  • Monitor
    • Monitor
    • Incidents
  • Packages & Registries
    • Packages & Registries
    • Package Registry
    • Container Registry
    • Infrastructure Registry
  • Analytics
    • Analytics
    • Value stream
    • CI/CD
    • Code review
    • Insights
    • Issue
    • Repository
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Activity
  • Graph
  • Create a new issue
  • Jobs
  • Commits
  • Issue Boards
Collapse sidebar
  • Open Subsurface Data Universe Software
  • Platform
  • Data Flow
  • Data Ingestion
  • External Data Sources
  • External Data Framework
  • Issues
  • #52

Closed
Open
Created Dec 20, 2020 by Debasis Chatterjee@debasiscOwner

Options for Consumer to manage bulk data files (ex: LAS, DLIS, SegY) from Data Provider

Refer to recent meeting held with @jrougeau and @Rawaa on this subject.

The current approach is to bring meta data of work-product-component (ex: WellLog) permanently across by fetch-and-ingest process. But for actual bulk data, the understanding is that Operator BigOil will bring the file across "on demand" later which will retain the file in "cache" for limited period (ex: 7 days). So, "Proxy delivery" process will have the intelligence to check if it is already in cache, or bring a fresh copy if it is not in cache.

The argument in favour of this is that Operator BigOil will always get a fresh copy of bulk data in case Subscription company made some updates at their end. In addition, this approach is more secured and automatically mitigates the risk of Operator Bigoil retaining subscribed data beyond license expiry.

This approach makes OSDU data in BigOil repository somewhat heterogeneous i.e. some data (loaded by BigOil's own Data Loaders by using standard OSDU ingestion process) appears in one way and some other subscribed data (integrated using EDS' feature of fetch-and-ingest).

We solicit input from other Operators about their expectation of handling of bulk data (ex: LAS, DLIS, SegY files). Copying to @Rawaa and @jyc9999 for their respective inputs.

Also copying to @nhung.nguyenparker for her input from Data Subscriber perspective (i.e. what is real life situation with her company's subscription data).

Edited Jun 02, 2021 by Debasis Chatterjee
Assignee
Assign to
Time tracking