OK so we've heard the stories of people using dimensions with 1m or more elements in them, but realistically how friendly are they to use in a production environment?
I’m testing a model right now that uses an 11 dim cube and I’m expanding one dim from around 26,000 elements to near 300,000. It’s a requirement to test if we can use our existing models as a template for various areas to gather the data we require but at a much finer level of detail. If successful it will replace various bespoke, non-supported systems using Excel or Access with a single, does the same for everyone TM1 system, and reduces their workload significantly.
My initial test failed on my 32bit dev server (as I knew it would) when I tried to load some dummy data and it broke the memory limit. My next test is to break up the models in to programmes and load less data, so this time I’m trying to cut down the large dim so it only has data for one of the programmes rather than all.
I cannot edit the dim via XDI because Excel is unable to handle something that big. So I turned off the properties pane and went to edit the dim via server explorer.
That was nearly and hour ago and it’s still thinking about it. It can’t be possible to run a production system with a dim this kind of size without getting totally naffed off waiting for it to do something. I know I can edit the dim with TI but having seen how badly this is behaving right now, I’ve grave concerns about its usability even if we moved to 64bit.
Does anyone seriously use a model this kind of size with any level of success?
Realistically Using Large Dimensions
- Steve Vincent
- Site Admin
- Posts: 1054
- Joined: Mon May 12, 2008 8:33 am
- OLAP Product: TM1
- Version: 10.2.2 FP1
- Excel Version: 2010
- Location: UK
Realistically Using Large Dimensions
If this were a dictatorship, it would be a heck of a lot easier, just so long as I'm the dictator.
Production: Planning Analytics 64 bit 2.0.5, Windows 2016 Server. Excel 2016, IE11 for t'internet
Production: Planning Analytics 64 bit 2.0.5, Windows 2016 Server. Excel 2016, IE11 for t'internet
-
- MVP
- Posts: 214
- Joined: Tue Nov 11, 2008 11:57 pm
- OLAP Product: TM1, CX
- Version: TM1 7x 8x 9x 10x CX 9.5 10.1
- Excel Version: XP 2003 2007 2010
- Location: Hungary
Re: Realistically Using Large Dimensions
Hello!I cannot edit the dim via XDI because Excel is unable to handle something that big. So I turned off the properties pane and went to edit the dim via server explorer.
That was nearly and hour ago and it’s still thinking about it. It can’t be possible to run a production system with a dim this kind of size without getting totally naffed off waiting for it to do something. I know I can edit the dim with TI but having seen how badly this is behaving right now, I’ve grave concerns about its usability even if we moved to 64bit.
What kind of TI problem you are referring in connection with dimension maintenance?
Regards,
Peter
Best Regards,
Peter
Peter
- Steve Vincent
- Site Admin
- Posts: 1054
- Joined: Mon May 12, 2008 8:33 am
- OLAP Product: TM1
- Version: 10.2.2 FP1
- Excel Version: 2010
- Location: UK
Re: Realistically Using Large Dimensions
Sorry, badly worded. I don't have an issue with a TI, i have an issue with the normal way i'd maintain a dimension (via the server explorer). In fact i could strip out all the stuff i didn't want in my test using TI and it ran in about 1 sec, so why the blazes does it take over an hour to manually select some elements then press the "delete" key? I know John has expressed the same concerns with populating the GUI, especially around attributes and the properties pane, but i ensured i had that turned off before i even started.
I get the impression that anything done via the GUI is executed in the background one item at a time. So if i selected 75,000 elements and hit delete, it would do them one at a time, recalculating any cubes as it went. I can't think of any other reason as to why the GUI would take so long and TI would be so quick, as TI would lock the whole model until it'd finished, thus only calculating once rather than 1000's of times.
Its only a guess, but this alone is enough for me to suggest we keep the models as small as we can, despite the issues with trying to maintain so many models and keeping their dims in sync.
I get the impression that anything done via the GUI is executed in the background one item at a time. So if i selected 75,000 elements and hit delete, it would do them one at a time, recalculating any cubes as it went. I can't think of any other reason as to why the GUI would take so long and TI would be so quick, as TI would lock the whole model until it'd finished, thus only calculating once rather than 1000's of times.
Its only a guess, but this alone is enough for me to suggest we keep the models as small as we can, despite the issues with trying to maintain so many models and keeping their dims in sync.
If this were a dictatorship, it would be a heck of a lot easier, just so long as I'm the dictator.
Production: Planning Analytics 64 bit 2.0.5, Windows 2016 Server. Excel 2016, IE11 for t'internet
Production: Planning Analytics 64 bit 2.0.5, Windows 2016 Server. Excel 2016, IE11 for t'internet
- Steve Rowe
- Site Admin
- Posts: 2455
- Joined: Wed May 14, 2008 4:25 pm
- OLAP Product: TM1
- Version: TM1 v6,v7,v8,v9,v10,v11+PAW
- Excel Version: Nearly all of them
Re: Realistically Using Large Dimensions
Hi Steve,
I don't personally work with the dimension editors or xdi at all when looking after dimensions. I don't have any experience of working with such a large dimension either.
Nearly all the dimensions in the systems I look after are maintained in an Excel table, that is then processed with TI into the system. I've found it much easier than the tools provided with TM1.
For a dimension the size yout talking about you'd need to move to having the standing metadata structure in SQL and read that into TM1 via TI. I would have thought it would be pretty straight forward...
The real problem (again no direct experience) is not how you would maintain the dimension but how the users would be able to use it, especially if you are looking to produce detailed xl worksheets.
Cheers,
I don't personally work with the dimension editors or xdi at all when looking after dimensions. I don't have any experience of working with such a large dimension either.
Nearly all the dimensions in the systems I look after are maintained in an Excel table, that is then processed with TI into the system. I've found it much easier than the tools provided with TM1.
For a dimension the size yout talking about you'd need to move to having the standing metadata structure in SQL and read that into TM1 via TI. I would have thought it would be pretty straight forward...
The real problem (again no direct experience) is not how you would maintain the dimension but how the users would be able to use it, especially if you are looking to produce detailed xl worksheets.
Cheers,
Technical Director
www.infocat.co.uk
www.infocat.co.uk
- Harvey
- Community Contributor
- Posts: 236
- Joined: Mon Aug 04, 2008 4:43 am
- OLAP Product: PA, TM1, CX, Palo
- Version: TM1 8.3 onwards
- Excel Version: 2003 onwards
- Contact:
Re: Realistically Using Large Dimensions
Hey Steve,
I don't want to be presumptuous, but I doubt you'd ever edit a dimension of 300,000 elements manually, either via XDI or Server Explorer.
The company I work with now was part of a project that includes a dimension with millions of elements (prepaid mobile phone numbers) and it is still working in production now, processing hundreds of thousands of transactions a day. I believe it is still considered one of the, if not THE, largest TM1 model in production.
I didn't work on it myself, but I could get some stats for you if you're interested in the details.
Cheers,
Harvey.
I don't want to be presumptuous, but I doubt you'd ever edit a dimension of 300,000 elements manually, either via XDI or Server Explorer.
The company I work with now was part of a project that includes a dimension with millions of elements (prepaid mobile phone numbers) and it is still working in production now, processing hundreds of thousands of transactions a day. I believe it is still considered one of the, if not THE, largest TM1 model in production.
I didn't work on it myself, but I could get some stats for you if you're interested in the details.
Cheers,
Harvey.
Take your TM1 experience to the next level - TM1Innovators.net
- jim wood
- Site Admin
- Posts: 3958
- Joined: Wed May 14, 2008 1:51 pm
- OLAP Product: TM1
- Version: PA 2.0.7
- Excel Version: Office 365
- Location: 37 East 18th Street New York
- Contact:
Re: Realistically Using Large Dimensions
Steve,
We have a 11 dimension cube with one dimension that has around 489k elements within it. It is used as a sandbox cube for consolidating data only. The memory usage is quite high. It currently has roughly 2 years worth of data within it and it takes 10gb. I have pulled data out of it and the respone time when pulling data at a hig level is really slow. Also when expanding the said large dimension there can be quite a bit of lag. If you keep the data sets small the performance isn't bad. (The fact that our Solaris box has really slow CPU's may have something to do with it.)
One other point. We have no rules on the cube to avoid us blowing our memory limit. (Our server memory usage is capped at 40gb)
I hope this helps,
Jim.
We have a 11 dimension cube with one dimension that has around 489k elements within it. It is used as a sandbox cube for consolidating data only. The memory usage is quite high. It currently has roughly 2 years worth of data within it and it takes 10gb. I have pulled data out of it and the respone time when pulling data at a hig level is really slow. Also when expanding the said large dimension there can be quite a bit of lag. If you keep the data sets small the performance isn't bad. (The fact that our Solaris box has really slow CPU's may have something to do with it.)
One other point. We have no rules on the cube to avoid us blowing our memory limit. (Our server memory usage is capped at 40gb)
I hope this helps,
Jim.
Struggling through the quagmire of life to reach the other side of who knows where.
Shop at Amazon
Jimbo PC Builds on YouTube
OS: Mac OS 11 PA Version: 2.0.7
Shop at Amazon
Jimbo PC Builds on YouTube
OS: Mac OS 11 PA Version: 2.0.7
- Steve Vincent
- Site Admin
- Posts: 1054
- Joined: Mon May 12, 2008 8:33 am
- OLAP Product: TM1
- Version: 10.2.2 FP1
- Excel Version: 2010
- Location: UK
Re: Realistically Using Large Dimensions
Thanks for the comments, all very helpful. I'm only going on my gut feeling of what people will want in this type of model (trying to pre-empt questions for when they do arise) but i am worried about response times. Editing the dim will mainly be done with TI (i hope) but this particular dim is made up from about 7 different csv files currently. The expanded version is likely to require about 10 seperate sources to build the total hierarchy. The problem that may arise is that some of this data *may* need to be manually added, due to the way the business will define its requirements. It might not, i could possibly work around it but either way it's all stuff to consider.
SQL is sadly not an option. The people who will ultimately be left to look after these models will not be IT literate to that level. Also, we don't have any SQL data sources they would be allowed to administer, apart from the overly complex Oracle Apex which i wish to avoid like the plague. Jim, does your large dimension have any hierarchy or is it fairly flat? One thing i might consider is ditching all the levels and try to keep it as flat as possible, i will need to write rules on this beast so the fewer # of datapoints it has the better.
The data should be sparse enough for it to work, but i won't know until i try. Speed really will be the key tho - if its massively slow then they'll get turned off by it and we'll end up with a nasty big failure on our hands. Thats an avoid-at-all-costs issue, even if it means not doing it at all...
SQL is sadly not an option. The people who will ultimately be left to look after these models will not be IT literate to that level. Also, we don't have any SQL data sources they would be allowed to administer, apart from the overly complex Oracle Apex which i wish to avoid like the plague. Jim, does your large dimension have any hierarchy or is it fairly flat? One thing i might consider is ditching all the levels and try to keep it as flat as possible, i will need to write rules on this beast so the fewer # of datapoints it has the better.
The data should be sparse enough for it to work, but i won't know until i try. Speed really will be the key tho - if its massively slow then they'll get turned off by it and we'll end up with a nasty big failure on our hands. Thats an avoid-at-all-costs issue, even if it means not doing it at all...
If this were a dictatorship, it would be a heck of a lot easier, just so long as I'm the dictator.
Production: Planning Analytics 64 bit 2.0.5, Windows 2016 Server. Excel 2016, IE11 for t'internet
Production: Planning Analytics 64 bit 2.0.5, Windows 2016 Server. Excel 2016, IE11 for t'internet
-
- Regular Participant
- Posts: 152
- Joined: Fri May 23, 2008 12:08 am
- OLAP Product: TM1 CX
- Version: 9.5 9.4.1 9.1.4 9.0 8.4
- Excel Version: 2003 2007
- Location: Melbourne, Australia
- Contact:
Re: Realistically Using Large Dimensions
Hi Steve,
Sorry if this is a bit late just haven't had time to post anything lately.
The largest dims we have in production would be a couple of SKU dims used in sales cubes for some retail clients with somewhere between 100K - 200K elements. Once consolidations are accounted for probably around double this in terms of the total elements. No real problems maintaining these or working with these. Recently I have done a couple of POCs replacing Powerplay with TM1 because Powerplay couldn't handle the dimension and cube size (very slow performance) - these were for dims with about 400K - 600K elements. TM1 handled it pretty well
The only way to maintain or edit dimensions of this size is with TI. It would be folly to ever attempt to manually edit a large dimension. IMO dimension editor struggles even with moderately sized dimensions of more than 1000 elements. The other big watch out is that with large dimensions the user base needs to be better educated as you have to strictly enforce a rule of "the properties pane must be off" when browsing a large dim.
Performance when calculating views generally good, but watch out for rules if you have a lot of dimensionality. Summary cubes for most reporting and drilling back to the mega cube when required helps a lot, as does splitting out the large dim into smaller summary dims (if possible depending on the model.) Optimising dimension order in cubes using large dims can also have a huge impact on performance.
Sorry if this is a bit late just haven't had time to post anything lately.
The largest dims we have in production would be a couple of SKU dims used in sales cubes for some retail clients with somewhere between 100K - 200K elements. Once consolidations are accounted for probably around double this in terms of the total elements. No real problems maintaining these or working with these. Recently I have done a couple of POCs replacing Powerplay with TM1 because Powerplay couldn't handle the dimension and cube size (very slow performance) - these were for dims with about 400K - 600K elements. TM1 handled it pretty well
The only way to maintain or edit dimensions of this size is with TI. It would be folly to ever attempt to manually edit a large dimension. IMO dimension editor struggles even with moderately sized dimensions of more than 1000 elements. The other big watch out is that with large dimensions the user base needs to be better educated as you have to strictly enforce a rule of "the properties pane must be off" when browsing a large dim.
Performance when calculating views generally good, but watch out for rules if you have a lot of dimensionality. Summary cubes for most reporting and drilling back to the mega cube when required helps a lot, as does splitting out the large dim into smaller summary dims (if possible depending on the model.) Optimising dimension order in cubes using large dims can also have a huge impact on performance.