New features delivered in 23.12
The following features and enhancements were included as part of 23.12.
Data Bridge 23.12 release includes the Data Bridge 3.7.0 release.
Creating publishing targets and mapping them to views
This release introduces a new concept of Publishing Targets that enables a more flexible and intuitive way to create CRM data event streaming and snapshot publishing configurations, enhances the user experience and paves the way for future strategic development of Data Bridge product. Data Bridge user interface has been refactored to support configuration of Publishing Targets and mapping them to Views (DATABRIDGE-1181). This new functionality enables you to:
Add publishing targets (data consumers), which determine where the data is published, on the new Publishing Targets tab on the Integrated Systems workspace (DATABRIDGE-1015 and DATABRIDGE-1260). Settings specific to each publishing target can now be configured on the Publishing Targets tab. Such configuration was previously found on the Global Settings workspace or under each View.
Configure multiple data publishing pipelines per view. This can now be achieved by adding one or more publishing targets to a view in the Publishing Target Mapping section in the Publishing Options tab in the View Builder, and configure the targets to enable snapshot exports, real-time streaming, and scheduling (DATABRIDGE-1171 and DATABRIDGE-1267).
Monitor the performance of real-time streaming and snapshot exports for each publishing target in the Monitoring Dashboard (DATABRIDGE-1175).
View the name of the publishing target and target type for errors in the Errors tab on the View Details page (DATABRIDGE-1176) and filter the errors data table by target type in the Errors tab (DATABRIDGE-1300).
Ingesting data from a Kafka topic
In addition to data publishing capabilities, this release of Data Bridge introduces the ability to ingest updates from a Kafka topic (DATABRIDGE-1140). The new Data Ingestion module enables:
Subscribing for event streaming updates from a Kafka topic. The Kafka topic is identified using the new Data Sources tab on the Integrated Systems workspace.
Publishing the data updates to a NexJ CRM or a third-party HTTP API. The publishing targets are defined on the Publishing Targets tab on the Integrated Systems workspace.
Support for Apache Avro de-serialization (DATABRIDGE-1599). Avro formatting can be specified when configuring a data source.
This major capability update greatly simplifies the task of integrating CRM with an Apache Kafka-based data hub, consuming data updates and de-serialization of Apache Avro messages into JSON. Data Ingestion configurations can be built by a Data Bridge administrator at run-time with the help of the Generic Views tab on the View Explorer workspace.
Support for non-persisted attributes in filters
It is now possible to configure whether a non-persisted attribute can be used as a view filter by specifying the filterable property on the attribute in the relevant Data Bridge subject area definition file (DATABRIDGE-1233). This feature is meant to be used at the discretion of the developer, with the non-persisted CRM model attributes that have query engine support for filtering.
For more information, see Using Data Bridge .