Andyblg's Blog

December 11, 2010

Analysis Services BI Semantic Model (Denali)

Filed under: Analysis Solution — andyblg @ 19:38
Tags: , , ,

At the PASS 2010 conference in Seattle MS announced the plans for Business Intelligence in SQL Server “Denali”. Note new Business Intelligence Semantic Model (BISM) in Analysis Services.

Quotation from technet blog (full article in Analysis Services – Roadmap for SQL Server “Denali” and Beyond)

The purpose of this article is to elaborate on the BI Semantic Model and how it compares to the existing BI models in the Microsoft BI stack, specifically the UDM (OLAP models) and SMDL (report models). Very simply put, the BI Semantic Model is a relational (tables and relationships) model with BI artifacts such as hierarchies and KPIs. It unifies the capabilities of SMDL models with many of the sophisticated BI semantics from the UDM. However it does not replace the UDM.

The BI Semantic Model can be viewed as a 3-layer model:

  • The Data Model layer that is exposed to client applications. Even though the BISM model is fundamentally relational, it can expose itself using both relational as well as multidimensional interfaces. OLAP-aware client applications such as Excel can consume the multidimensional interface and send MDX queries to the model. On the other hand, a free-form reporting application such as Crescent can use the relational interface and send DAX queries.
  • The Business Logic layer that encapsulates the intelligence in the model. The business logic is created by the model author using DAX (Data Analysis Expressions) or MDX (Multidimensional Expressions). DAX is an expression language based on Excel formulas that was introduced in PowerPivot and built on relational concepts. It does not offer the power and flexibility that MDX does, but it is simpler to use and requires minimal tuning. There will always be sophisticated BI applications that need the power of MDX calculations and we envision that the BI Semantic Model will offer the choice of MDX as well, but this will likely come in a release after Denali.
  • The Data Access layer that integrates data from various sources – relational databases, business applications, flat files, OData feeds, etc. There are two options for data access – cached and realtime. The cached mode pulls in data from all the sources and stores in the VertiPaq in-memory column store. VertiPaq is a breakthrough technology that encapsulates state-of-art data compression algorithms along with a sophisticated multi-threaded query processor that is optimized for the latest multi-core chipsets, thereby delivering blazing fast performance with no need for indexing, aggregates or tuning. The realtime mode, on the other hand, is a completely passthrough mode that pushes the query processing and business logic evaluation down to the data source, thereby exploiting the capabilities of the source system and avoiding the need to copy the data into the VertiPaq store. Obviously there is a tradeoff between these two modes (the high performance of VertiPaq versus the latency and overhead associated with copying the data into VertiPaq) and that choice is left to the model author


1 Comment »

  1. Andyblg…

    […] something about andyblg[…]…

    Trackback by Andyblg — January 9, 2011 @ 15:53 | Reply

RSS feed for comments on this post. TrackBack URI

Leave a Reply

Fill in your details below or click an icon to log in: Logo

You are commenting using your account. Log Out / Change )

Twitter picture

You are commenting using your Twitter account. Log Out / Change )

Facebook photo

You are commenting using your Facebook account. Log Out / Change )

Google+ photo

You are commenting using your Google+ account. Log Out / Change )

Connecting to %s

Blog at

%d bloggers like this: