Bookwire Aggregators
When we designed the Onix intake for Ingram it was build to ingest publications to Ingram's tenant, and share them through the marketplace.
When we designed the Onix intake for Ingram it was build to ingest publications to Ingram's tenant, and share them through the marketplace.
Configuration
The Content Import system is a browser-based bulk import dashboard that allows tenants to upload spreadsheets (with metadata) and content files (PDF, EPUB, audio) to create products in Farfalla. Users upload files directly to S3 via presigned URLs, provide a metadata spreadsheet that maps each row to a content file, and the system validates, batches, and sends everything to Farfalla's /api/v3/content/bulk endpoint.
Medusa has two content intake systems that share the ContentIntake model as their entry point. Each ContentIntake record has a type enum — either onix or import — which determines which processing pipeline handles its data. Both systems ultimately send content to Farfalla for ingestion.
This is a really general overview of the content intake flow, from a content source (Bookwire, Ingram) sending
Here we present the information that we collect from the ONIX files that are sent to us for storage in Publica.la. We will indicate the type of data received, its relationship with the ONIX standard, the extractor responsible for obtaining such information, the required data, and the information that will be stored.
- https://ns.editeur.org/onix/en
To receive Onix intakes, we had to provide an SFTP server. On this SFTP server, is where the content source (Bookwire, Ingram) sends all the content that they provide. The problem is that they send GBs of data.
Repository: gitlab.com/publicala/medusa
End-to-end flow of a content import, from file upload to completion.
The Content Import system supports two spreadsheet formats: Publica.la and VitalSource. Accepted file types are XLSX, XLS, and CSV (max 10 MB).
From wikipedia: