Published | March 19, 2018


Operationalize Business Results with Semantic Modelling

Operationalize Results with Semantic Modeling

How's your data lake looking?

The flexibility and cost effectiveness of Apache Hadoop was quickly recognized by many organizations as an effective delivery vehicle to empower business users with operational self-service query and analytic capabilities.  Many organizations established, or are presently establishing, data lakes for operational intelligence query and analytics capabilities for the field personnel who need them most, best understand the data, and are the most capable of actioning insights gleaned.  However, aligning the usable, credible, and high quality data elements available in the lake with other streaming feeds poses data quality challenges using traditional techniques.

Provisioning data for analytics processes usually begins with the data lake contents.  Combining, integrating or merging these results with data in motion whether it’s from a message bus, a streaming feed or intermittent latency service often creates data quality challenges.  Different and often disparate technologies need to unite content in a singular result, creating tremendous challenges for the organization to maintain a cohesive understanding and apply it correctly.

Let’s expand on the Customer Lifetime Value blog example and model the results, component by component.  Revenue is a great one to begin with. This will undoubtedly start with balances comprised of results batched into a data lake and then combined with messages from the bus or streaming micro-service feeds sourced internally, depending on the specific transaction underway.  Consider the computed revenue at this point as a sharable resource to other analytic projects.  The top percentile of this high revenue segment is interesting to analyze from a trend perspective to determine which traffic patterns created or fostered this result.  It’s also interesting from a monitoring perspective to identify when a customer is transacting more or less than usual.

From a semantic model perspective, we have within the domain of customer an attribute, specifically a computed measure called revenue.  Revenue can be considered to be initialized from a data at rest element housed in the data lake.  Revenue can also be mapped to messages pulled from the enterprise bus – these messages may specify specific transactions that occurred since the last refresh of the lake.  Revenue can also be considered as a mapping to an internal micro-service that contains transactions not available on the bus, usually third-party in nature.  Another computed revenue attribute, called the revenue percentile, is calculated from the domain of all mapped customers and their revenue – these could be computed within time-specified ranges to form meaningful variables such as year-to-date revenue, revenue since inception, or top percentile of customers ranked by revenue this year.

With a semantic attribute model, emphasis is now placed on understanding the attribute and its context.  Notice the complete lack of structural dependency, technical nomenclature (save the mapping), and the ability to apply the results in a meaningful setting.  The attribute can power historical analytics as well as alerting mechanisms and can service multiple consumers without the peril of additional scripting, loading, processing or technical mapping.  To learn more about semantic modelling, check out this on-line course and materials from the University of Florida, or follow me at NexJ Systems.

With NexJ DAi, an attribute model that natively unites data at rest or in motion and presents results using a consistent terminology helps firms to better provision results for both user query and analytic efforts.  With NexJ DAi, integration occurs at an attribute level and, once resolved, the attribute can be assigned to many views. Attributes can integrate content from streaming web services and messages or from data at rest like databases or files. Attributes can also define a computation, allowing for centralization and review of calculations, easily modified and tailored to address specific needs – all leveraging the existing attributes already defined.  The NexJ DAi engine will natively publish attributes to associated views so changes in operational systems, updates from streaming services or the latest messages are reflected.

Takeaways To Date

Hadoop EcoSystemMultiple modelling techniquesNexJ DAI provides a semantic model

Key AspectResponseNexJ DAI
Hadoop EcoSystemHDFS Low Cost Commodity Storage
Hive In-Place Query
NexJ DAI integrates with Hadoop data lakes as a potential source system
Hadoop EcoSystemSpark In-Memory QueryNexJ DAI provisions semantic view data consumable through a Spark Adapter
Hadoop Data LakeBatch LoadNexJ DAI provisions the most up to date data
Hadoop EcoSystemSeparate technology for streamingNexJ DAI addresses both data at rest and in motion

How does your organization use a data lake?  What 360-degree data views power your analytics?  We welcome your thoughts, value your insights and action your feedback: share below!

speaker_notes Post Comments

Author: Matthew Bogart

Vice President, Marketing

Matthew is responsible for building awareness for NexJ and demand for its solutions. He regularly engages with analysts, key industry stakeholders, and thought leaders to stay abreast of technology innovation and financial services industry trends and challenges.

Matthew will be sharing his insight and perspective on the enterprise customer management market and the issues affecting the financial services industry today in regular contributions to the NexJ blog. He encourages readers to take part in the conversation or reach out to him directly with their observations on market trends and issues.

Comments Off on Operationalize Business Results with Semantic Modelling

Comments are closed

Blog Categories

Blogs by Date