Quick clarification, are you looking to ‘load’ a resource description of the dataset, or the dataset itself?
If you are looking to load a research dataset, then my view is that while that is a totally valid FOLIO pursuit, it is out of scope of the Codex. I hope that someone will build that infrastructure on the FOLIO platform, and a suite of apps for it. @nassar has been pursuing that idea as a side project and has explored at least one possible direction one might take, but not the only way. FOLIO is a platform, so you’re never limited to just one way of doing things.
If you’re looking to load bibliographic descriptions of data set(s) into the Codex you have options.
The Codex grew out of a desire to have a common data model and interface that would span a lot of popular and emergent data models that already exists in the wild for the purpose of bibliographic description and electronic and print holdings management. Nothing more. It’s scope and capabilities are to my mind influenced by a few important factors
-
The functional needs driving the initial set of apps, which focus largely on core library administration functions. Specifically resource/inventory management.
-
The desire to try to harmonize or bring together print and electronic resource management into a single data model to facilitate a more integrated solution across the two main content delivery mechanisms
-
A desire to support multiple, detailed schema for description (including BIBFRAME) while allowing a set of relatively naive apps to ignore all but the data elements they need to support common workflows
-
A desire to support multiple storage mechanisms, including a conventional, local-to-the-system data well or multiple remote knowledge bases, union catalogs, authorities, or LOD data sources. These storage mechanisms can be facilitated simply by writing new implementations of the Codex interface.
The Codex leans on BIBFRAME in the sense that it recognizes a similar/identical set of entities. The “Instance” and “Item” are the one we focus on initially in this stage because you need them to drive present day functions. The core Codex Instance record is our most basic descriptive metadata set that the v1 apps can rely on (when we discussed this in Ireland, I suggested we call it the “Real Dublin Core”). If richer metadata is associated with a given Codex record, e.g. as a linked MARC or BIBFRAME or MODS record, then v1 apps can be extended to take advantage of that richer metadata if desired, or someone might write a new app specifically to take advantage of some specific format.
Back to your question
Assuming that you want to ‘load’ descriptive metadata, not actual research datasets: First, I’d say, do you really want to load it? Or, do you want to write an implementation of the Codex Instance interface that will expose the metadata to FOLIO apps while leaving it in place in SDL (or whatever). If you really want to load it, you have options…
Assume we have a “local-to-FOLIO” storage module for metadata which allows you to load up data in any schema you like provided there exists a crosswalk to the Codex (so that v1 apps can see the data without having to know your schema).
Is your data set ‘supported’? I.e. does a crosswalk already exist? If not, you either create the crosswalk, then load the data, or, you map your data into one of the already existing schemas like MARC, BIBFRAME, whatever. Then you load it. As a special case, obviously we can support simply loading the data in the Codex schema, which would imply a no-op crosswalk.
What if your data set can’t be captured as a set of ‘records’ because it’s a free-floating set of triples or a RDBMS. Then, you will have to create a ‘view’ of your data set that represent entities or units of descriptive metadata conceptually similar to records, and load them. If you can’t do THAT, then your data set probably won’t make sense to any of the apps written to deal with a Codex view of the world.
Again, the Codex is a ‘view’ of the universe of bibliographic and holding metadata designed to let us do a certain set of things… by all means we should push the boundaries of what that is and what we can do with it, that’s the whole point… I keep expecting someone to make a pitch to use JSON-LD and shared schemas for the Codex and I’d be really intrigued by that conversation. But if you want to build apps in FOLIO that lean on all the expressive power of the semantic web and the breakdown of normal notions of entities/records, etc., then I don’t think the Codex is or should be the tool for that… trying to make it that in the short term would only trip us up.
At the same time, I really do hope that someone will integrate a triplestore into FOLIO and start to build apps that leverage it to do library and non-library things. It’s just that for the part of FOLIO which is some people racing to build a suite of apps you can use to run a library, that’s not the approach that has been chosen.
Please note, I’m not trying to shut down a conversation about Linked Data and where FOLIO should be moving. We need that conversation. But that’s a really broad conversation, and, there are actually some parts of that conversation that probably SHOULD inform even the present roadmap, while others can help us keep an eye on the long game even while we focus on something less ambitious right this moment.