Posted by Alex Kuznetsov on 12/08/06 03:23
meeraguna@gmail.com wrote:
> We have 20 -30 normalized tables in our dartabase . Also we have 4
> tables where we store the calculated data fron those normalised tables.
> The Reason we have these 4 denormalised tables is when we try to do
> the calcultion on the fly, our site becomes very slow. So We have
> precalculated and stored it in 4 tables.
>
> The Process we use to do the precalcultion, will get do the
> calculation and and store it in a temp table. It will compare the the
> temp with denormalised tables and insert new rows , delte the old one
> ans update if any changes.This process take about 20 mins - 60mins. It
> takes long time because in this process we first do the calculation
> regardless of changes and then do a compare to see what are changed and
> remove if any rows are deleted, and insert new rowsand update the
> changes.
>
>
> Now we like to capture the rows/columns changed in the normalised
> tables and do only those chages to the denormalised table , which we
> are hoping will reduce the processing time by atleast 50%
>
>
> WE have upgraded to SQL SERVER 2005.So We like to use the new
> technology for this process.
> I have to design the a model to capture the changes and updated only
> those changes.
>
> I have the list of normalised tables and te columns which will affect
> the end results.
>
>
> I thought of using Triggers or OUTPUT clause to capture the changes.
> Please help me with the any ideas how to design the new process
We do not do much of this denormalization stuff manually these days -
covering indexes and indexed views usually boost performance well
enough and they do it automatically...
[Back to original message]
|