Complex expression to remove currency symbols?
Hi all, we have a process to format data in the csv before it's imported into OneStream. This seems like extra work and I want to create a complex expression in the Data Source to remove this step. My excel formula is =VALUE(SUBSTITUTE(SUBSTITUTE(H9,"¥",""),",","")).... all the way down for column H or the Amt field import. I've tried putting this in the complex expression but I do not have the dimensions defined properly.... does anyone have an example or suggestions they can point me towards? thank you!!!24Views0likes1CommentXFGETCELL(), XFSETCELL() PERFORMANCE
Hi all, Every forecast cycle we get the same performance complaints from the same region group of users. They are always uploading massive amounts of data and refreshing massive amounts of data in the same workbooks using the XF functionality in Excel. To be clear, we have tried multiple times to convince them that they should be importing that data via CSV and advising that they will not get the same performance using the XF functions. We have also tried having OneStream re-index the databases before these cycles and we have regular daily app pool recycles. The OneStream data logs really don't show any significant refresh issues (less than 5 minutes usually) but the one user doing all this work is insisting that it's taking hours and hours to finish refreshing and they often have to close and reopen Excel and re-launch OneStream. We have raised multiple tickets and OneStream does not see any back end issues. We have looked at their Excel preferences and tried different options there but no improvement. I am 100% out of ideas so if anyone out there has any thoughts, would love to hear what you've done to resolve these issues. I'm going to post this as well somewhere in the community where Excel discussions happen. Thanks for any advice.20Views0likes1CommentCubeview Drill Down issue: Load Results for Importe Cells is empty
Hello there, We are having an issue when using the Drill Down option from a Cubeview, and it's that after getting to a base data cell within the Drill Down window, and then trying to navigate to stage using the "Load Results for Imported Cell", the results window is displayed empty: The workflow that is being used it's a simple Import, Validate and Load, that retrieves the imported data from a Connector and there is no further data processing once it's in the Cube. From the Workflow Page we can use Drill Down and Drill Back just fine. Any ideas of why is this happening? Thank you. OriolSolved40Views0likes4CommentsValidation on Data Source: Restrict Entities to _PX Suffix and Metadata Text Contains PSC^
Hi Community, I’m working on a data load in OneStream and need to enforce two validations on the Entity dimension during the upload process: Entity name must end with _PX. Entity’s metadata Text field (e.g., Text1/Text2) must contain PSC^. (Basically Prevent data from being loaded to anything but a PSC entity) Currently, I see the Logical Expression and Override Settings on the Data Source column, but it seems limited to simple operators like Ends With or Like. My question is: Can this be achieved using a complex expression directly in the Data Source column settings? Or do I need to implement this as a Transformation Validate rule or a Data Source Business Rule? If a Business Rule is the right approach, could someone share a best-practice snippet for checking both conditions (including metadata lookup)? Should I create a conditional rule and add to the logical expression above?30Views0likes2CommentsTime/data setting in data source/Transformation rule for Multiperiod upload (delimited files)
Hi Everyone, I'm trying to build Data source and transformation rule for multi period manual data load through single file. The import step didn't succussed. Can anyone suggest time settings in data source & transformation rule. I selected data type as "Stored Data key text" in Data source . My source file format is Col1 has GL accounts, Col2-13 Jan to Dec data.Solved59Views0likes5CommentsMatrix Data Load Basics
I have a matrix data load question. I get the concept of how the matrix data load works. I haven't done too many of them, so I want to make sure I'm not missing the basics. The file I have has the year on one of the first lines in column 9. The next row contains the headers with periods in 12 separate columns. Is there an easy way to get the year from the first line (without a lot of code)? Or should I get year based on the POV? Then I am guessing a complex expression in the time columns to build the OS time key? I know I can do all of this in code, but I don't want to skip over any out of the box functionality or best practices. Any good guides on some of the best approaches with the Matrix load? Thanks, ScottSolved1.2KViews0likes5CommentsHow do you know if you are using the Roslyn compiler?
Hotfix 8.4.3 addresses, among other things, security vulnerabilities found in the third-party Roslyn Compiler DLL. How do you know if you are using the Roslyn compiler in your implementation of custom code and, therefore, at risk if you are? Is the Roslyn Compiler only used by the WinSCP libraries (and therefore its removal is WinSCP-related)? Or are there instances where Smart Integration Functions would be using the Roslyn compiler outside of the use of WinSCP?SolvedWorkflow channels with scenario copies
Hi All, In our multi year planning process, most entities need to upload four sets of fixed costs (all loaded by the same team, but sent to that entity's finance team by different functional teams), in addition to all the driver-based costs we're calculating (mostly based on form inputs). For the four sets of fixed costs, we have four import profiles under the same parent workflow. The entities import their files and then run a DM sequence from a dashboard to run the validate/load for a range of years they can select. When they run this sequence, all import profiles (under the same parent) that have been imported are validated and loaded. We had not set up workflow channels on the different loads, and everything was working as expected. The different import profiles will use some of the same accounts, but different sets of cost centers. We did not see any collision of data when the same accounts were loaded, and if a balance that had previously been loaded in one import was removed from the import template and reimported, the data in that account would be cleared from the cube (even if that account was only used by one of the four imports) after the next validate/load. Now that we have started to copy cube data from one completed scenario (CY) to another blank scenario (CY_v2), so that some changes can be made to CY_v2, we were running into issues with our fixed cost loads in CY_v2 (balances not clearing, other values being overwritten). This makes sense, and we had not considered loading the fixed costs to copied scenarios. We would like to be able to only load one of the four imports (if say the G&A load is the only one with changes) in the CY_v2 scenario. We were hoping that workflow channels on the imports would solve the issue in the copied-to scenarios, but after completing that set up (changing accounts to nodatalock, setting up different data source members for each of the four imports, creating the channels, and applying the channels to the data source members), we are still seeing balances "stuck" in CY_v2 when previously loaded balances to a unique account are removed from the latest upload for one of the imports. We had also gone back and reloaded the "source" scenario with the channels applied, before copying the data and trying to load only a single template to CY_v2. Is there a way to use workflow channels so that all data loaded through that channel is cleared and replaced when a new import is loaded to a data source member where the channel is applied, or will accounts no longer in the latest import be left un-cleared? The documentation on the level 2 and level 3 data units seems to indicate that those accounts will not be cleared, but that does not match the behavior we see in the original loads of multiple sibling import profiles (described and bolded in my second paragraph). It also doesn't seem to match how a typical, iterative planning process would work. Any feedback would be appreciated. Thanks, James41Views0likes0CommentsTransformation Rules - Possible Bug
I have an issue, since we migrated to V9, OS is adding on its own a space after transforming a member and this prevent us from loading the data to the member. (This is in the validation step when it fails) Data source has been checked, adjusted, and tested in many ways. The extra space doesn't come from here. Transformation rules have also been tested in many ways, and the space is also not here, we event tried one-to-one, masks, all ,and this is not the issue The error is only happening with 2 UD3 members US4 & CA4 both ending in 4, I created a fake one for testing "US41" and this one works fine. I have been checking with OS support, but they just want to repeat over and over the same testing we already did with them on a call To me, after checking all that I could, seems like a bug in the new V9.1, but if anyone has any ideas please let me know.