Transformer architecture

Here’s the internal architecture of Caplin Transformer:

Transformer’s internal architecture

Transformer’s a DataSource application, just like Liberator and the Integration Adapters deployed in the Caplin Platform.

  • It receives data from other DataSource applications, typically:

    • Integration Adapters (for example, a Pricing Adapter that feeds Transformer with raw price information),

    • other Transformers (for example, when you have a lot of data transformations to make and you want to split them between Transformers in addition to splitting them between modules within a Transformer),

    • or even a Liberator that publishes data supplied by clients (see Client publication).

  • The Data sink element of Transformer extracts the incoming data, via the DataSource API.

  • The Central processing element sends the data for each incoming subject to the relevant Transformer module(s), which execute the specific algorithms that transform the data. Central processing also maintains a cache of the data.

  • The amended data is either output to one or more external applications, such as databases used to store historical records, or it’s sent back to the Transformer core.

  • Once back in the core, the Data source element sends the amended data out, via the DataSource API, to other subscribing DataSource applications. These would typically be:

    • Liberator (which pushes the data through to client applications via StreamLink),

    • other Integration Adapters (which would publish the amended data to an external system),

    • or even other Transformers.

See also: