Recent Discussions
Import with Vertical Extensibility or Extensible Dimensions
Hi, I can’t get my data import to work with vertical extensibility, and I’d like to understand how this situation is supposed to be handled. In short, the import validation fails on accounts that are extended for an entity. The entity belongs to a cube that is vertically extended from the primary workflow cube. Detailed situation I have extended dimensions and two cubes that use those dimensions. CubeParent contains a parent entity with dimensions applied. The parent entity resides in the parent entity dimension. CubeChild has dimensions that are extended from the parent’s dimensions. The child entity resides in the child entity dimension. As a result, each cube has its own unique entity dimension containing its respective entity. Additional configuration details: CubeChild has Is Top Level Cube For Workflow set to False. CubeParent is the top-level cube and has the Workflow Profile created from it. The child entity has been added as a child of the parent entity in the parent’s entity dimension (see screenshot). CubeParent has the Cube References tab configured correctly. Data import issue I have configured a data source that reads from a CSV file. The data source successfully loads the CSV into the staging area. However, the transformation step fails. The failure occurs only for the child entity’s extended account, not accounts in the ParentCube account dimension. The transformation rules when created are associated with a cube. If the cube is not that of the workflow profile the transformation rule does not appear in the import configuration step. Question How should data be imported for an entity that exists in a workflow that has different dimensional detail than the cube it belongs to because it is using vertical extensibility?Montreux1 hour agoNew Contributor III5Views0likes0CommentsData Import Step - Load validated records
We are looking for a load into a financial cube. Mappings are already done in the source system. Is there a way to load records that pass the validation and write the other records to an error protocol. The idea would be that not the whole load is rejected but only records that contain errors, those being written to an error protocol. Many thanks for your supportjuergzbinden23 hours agoNew Contributor II25Views0likes2CommentsAdd Reminder for Flip Sign Flag on Account Mapping During Load
Our income, liability, equity accounts need to have the flip sign flag checked but our users always forget. Is there a way to add a pop up when they first get to this screen (after they save it is too late).mgreenberg1 day agoContributor II27Views0likes2CommentsMapping IC based if account is IC
We want to set the IC value to none if an account is not IC. We are currently doing this in the IC dimension in the datasource and looking up the account, getting the target and then checking If IC. If it is not IC we pass None as the source value. This works fine unless they upload a new account because it has not been mapped yet it doesnt know its IC and maps to None. The user has to reupload the same file after they map the account so that they get an IC value in the source. I was thinking that I could put a BR on the IC transformation rule but the problem is it reads the 1:1 (where I have the None->None) before it gets to the section for BRs. I guess I could remove the 1:1 None rule and then in the BR check to see if the account is IC and if so then get the IC target. I am just curious if there is a better way to handle this that I have not thought of. We have a similar issue as we store function in a user field but we only want it populated for specific expense accounts otherwise it should be none. When they have a new expense account we have the same problem that they have to reload the import file after they map the expense account.mgreenberg1 day agoContributor II36Views0likes2CommentsNo valid DataKeys (Scenario / Time) found in data source.
Hello everyone, I'm trying to create my first import from scratch. Unfortunately, it always gives me the same error : No valid DataKeys (Scenario / Time) found in data source. Review the source data load processing log and check one-to-one transformation rules to ensure that you have created proper Scenario and Time dimension rules for this data source. And no line is imported or displayed in the logs. However, I've gone down each parameter and everything seems correct and consistent with what I have on other import files. Could you help me investigate? My settings : All the UD1 to UD8 are on None. And the log :SolvedJérémyRenard2 days agoNew Contributor355Views0likes9CommentsConnect MS Fabric Lakehouse?
Hi all, I'm currently exploring ways to extract data from a Microsoft Fabric Lakehouse into OneStream. The obvious way would be to connect to the SQL endpoint just like any other DB. In this case, what would be the best authentication method? I've been testing it out with my own user on MS SQL Server Management Studio using MFA, and I have no idea what the best way would be to create credentials for the connection string in OneStream. In case connecting to the SQL endpoint with a connection string isn't a good idea, what would be a better practice? Thanks.mireles6 days agoNew Contributor III39Views1like1CommentErrors creating New Task in Data Import Schedule Manager
I am getting several errors when setting up a New Task in the Data Import Schedule Manager. We are on OS version 9.01.17403, and I have installed the latest version of DSM, which is 8.4.0_SV100. I suspect this may be a version compatibility issue, so I am curious if anyone has been able to get this solution to work in a 9.01 application. I have already uninstalled and reinstalled the solution, which didn’t resolve the issues. Below are the two errors I am seeing: When choosing Global Scenario from the Scenario(s) drop down list, I get an immediate error “Error processing member. The item was not found. Member, 11111111.” The details state: Unable to execute Business Rule ‘DSM_Paramhelper’ where it appears to be trying to call the Global Scenario by OneStream.Client.Api.DashboardsAjaxServiceReference.DashboardsAjaxServiceClient.GetParameterDisplayInfosUsingDashboardNameCompressed(SessionInfo si, LoadDashboardInfo loadDashboardInfo, Boolean isForDashboardUIWithInteractiveComponents, Dictionary`2 custSubstVarsAlreadyResolved). If I pick a specific Scenario, I am getting a different error. It allows me to pick the Scenario and Time, but when save the Task, I get “The input string ‘’ was not in a correct format.”. The error details show it is an issue with the same Business Rule ‘DSM_SolutionHelper’ where Conversion from string “” to type ‘Double’ is not valid. The input string ‘’ was not in a correct format. OneStream.Client.Api.DashboardsAjaxServiceReference.DashboardsAjaxServiceClient.StartExecuteSelectionChangedServerTaskCompressed(SessionInfo si, Boolean isSystemLevel, Guid primaryDashboardID, Guid embeddedDashboardID, Guid componentID, PageInstanceInfo pageInstanceInfo, XFSelectionChangedServerTaskInfo serverTaskInfo) Any advice on how to correct these issues would be greatly appreciated.86Views1like1CommentInput string was not in a correct format - V8 upgrade (Marketplace Solutions)
Hi Team, As part of V8 upgrade, I have uninstalled UI. and then loaded SV102 zip file for capital planning. But receive the following message. Any one experienced this and resolved?SolvedChadwin13 days agoNew Contributor673Views0likes3CommentsExport Cubeview with only the columns defined
Hi All, I am trying to export a cubeview using FdxExecuteCubeView, but I get every dimension even though my cubeview only shows 4 dimensions. What is the best way to export a cubeview to retain its view. I want to export multiple cube views and load them to a table so that I can have all of the cubeview data exported through one Data Adapter call via the rest API. Right now, I export one view I get 3 million rows, but when I aggregate it to the level in the cubeview I end up with 28,000 rows. I'd like to avoid exporting 12 million rows if possible. Thanks, Scottssmith-nova15 days agoContributor31Views0likes2CommentsHow to write Smart Integration Function in Assembly
Hi, I have created one one Smart integration function in to transfer file from sftp server to share point server. can we use smart integration function as a file or use existing smart integration function as dependency , not sure just need suggestions.rhm18 days agoNew Contributor59Views0likes1Comment