Import with Vertical Extensibility or Extensible Dimensions
Hi, I can’t get my data import to work with vertical extensibility, and I’d like to understand how this situation is supposed to be handled. In short, the import validation fails on accounts that are extended for an entity. The entity belongs to a cube that is vertically extended from the primary workflow cube. Detailed situation I have extended dimensions and two cubes that use those dimensions. CubeParent contains a parent entity with dimensions applied. The parent entity resides in the parent entity dimension. CubeChild has dimensions that are extended from the parent’s dimensions. The child entity resides in the child entity dimension. As a result, each cube has its own unique entity dimension containing its respective entity. Additional configuration details: CubeChild has Is Top Level Cube For Workflow set to False. CubeParent is the top-level cube and has the Workflow Profile created from it. The child entity has been added as a child of the parent entity in the parent’s entity dimension (see screenshot). CubeParent has the Cube References tab configured correctly. Data import issue I have configured a data source that reads from a CSV file. The data source successfully loads the CSV into the staging area. However, the transformation step fails. The failure occurs only for the child entity’s extended account, not accounts in the ParentCube account dimension. The transformation rules when created are associated with a cube. If the cube is not that of the workflow profile the transformation rule does not appear in the import configuration step. Question How should data be imported for an entity that exists in a workflow that has different dimensional detail than the cube it belongs to because it is using vertical extensibility?Solved51Views0likes4CommentsData Import Step - Load validated records
We are looking for a load into a financial cube. Mappings are already done in the source system. Is there a way to load records that pass the validation and write the other records to an error protocol. The idea would be that not the whole load is rejected but only records that contain errors, those being written to an error protocol. Many thanks for your support28Views0likes2CommentsConnect MS Fabric Lakehouse?
Hi all, I'm currently exploring ways to extract data from a Microsoft Fabric Lakehouse into OneStream. The obvious way would be to connect to the SQL endpoint just like any other DB. In this case, what would be the best authentication method? I've been testing it out with my own user on MS SQL Server Management Studio using MFA, and I have no idea what the best way would be to create credentials for the connection string in OneStream. In case connecting to the SQL endpoint with a connection string isn't a good idea, what would be a better practice? Thanks.Errors creating New Task in Data Import Schedule Manager
I am getting several errors when setting up a New Task in the Data Import Schedule Manager. We are on OS version 9.01.17403, and I have installed the latest version of DSM, which is 8.4.0_SV100. I suspect this may be a version compatibility issue, so I am curious if anyone has been able to get this solution to work in a 9.01 application. I have already uninstalled and reinstalled the solution, which didn’t resolve the issues. Below are the two errors I am seeing: When choosing Global Scenario from the Scenario(s) drop down list, I get an immediate error “Error processing member. The item was not found. Member, 11111111.” The details state: Unable to execute Business Rule ‘DSM_Paramhelper’ where it appears to be trying to call the Global Scenario by OneStream.Client.Api.DashboardsAjaxServiceReference.DashboardsAjaxServiceClient.GetParameterDisplayInfosUsingDashboardNameCompressed(SessionInfo si, LoadDashboardInfo loadDashboardInfo, Boolean isForDashboardUIWithInteractiveComponents, Dictionary`2 custSubstVarsAlreadyResolved). If I pick a specific Scenario, I am getting a different error. It allows me to pick the Scenario and Time, but when save the Task, I get “The input string ‘’ was not in a correct format.”. The error details show it is an issue with the same Business Rule ‘DSM_SolutionHelper’ where Conversion from string “” to type ‘Double’ is not valid. The input string ‘’ was not in a correct format. OneStream.Client.Api.DashboardsAjaxServiceReference.DashboardsAjaxServiceClient.StartExecuteSelectionChangedServerTaskCompressed(SessionInfo si, Boolean isSystemLevel, Guid primaryDashboardID, Guid embeddedDashboardID, Guid componentID, PageInstanceInfo pageInstanceInfo, XFSelectionChangedServerTaskInfo serverTaskInfo) Any advice on how to correct these issues would be greatly appreciated.Export Cubeview with only the columns defined
Hi All, I am trying to export a cubeview using FdxExecuteCubeView, but I get every dimension even though my cubeview only shows 4 dimensions. What is the best way to export a cubeview to retain its view. I want to export multiple cube views and load them to a table so that I can have all of the cubeview data exported through one Data Adapter call via the rest API. Right now, I export one view I get 3 million rows, but when I aggregate it to the level in the cubeview I end up with 28,000 rows. I'd like to avoid exporting 12 million rows if possible. Thanks, Scott37Views0likes2CommentsHow to write Smart Integration Function in Assembly
Hi, I have created one one Smart integration function in to transfer file from sftp server to share point server. can we use smart integration function as a file or use existing smart integration function as dependency , not sure just need suggestions.Solved60Views0likes1CommentFDXExecuteCubeView with Consol filter
Hi All, How would I go about passing a Consol member as part of the POV to a FDXExecuteCubeView call? I want to pull one extract that is in Local currency and another that is in USD. Can I do that with this call, or do I need to set up 2 cube views, one set to Local and another set to YTD? Thanks, ScottSolved54Views0likes3CommentsFdxExecuteCubeView parameters
If I build a cubeview that contains the desired POV and functions on its own, using FdxExecuteCubeView seems to work great. However, there is no documentation available on actually using the FDX functions and I can find no functional examples of how to pass the various filters/members from the FdxExecuteCubeView function to the cubeview it's actually calling. For example, I can enter T#2022M10 as part of the POV for the cube view and then FdxExecuteCubeView will create the table. However, if I pass the timeMemFilter a value of T#2022M10, I get no data. It seems obvious that I need to put something in the cube view POV to tell it to use this value but there's no indication of what that might be. Am I missing something obvious here? Is there some missing documentation or fundamental functionality that I'm unfamiliar with?Solved6.9KViews1like12CommentsEncryptText And DecryptText Replacement
Does anyone know what replaces the 2 utility commands below? I get a warning message that those commands are obsolete but it does not give any indication or hints of what the new commands are or what's currently available? BrApi.Utilities.EncryptText andBrApi.Utilities.DecryptText Any help is highly appreciated. Thanks, JunSolved171Views0likes3CommentsHow to check data load status of WFs with load overlapped siblings enabled
Hi, Is there a way to check status of all load overlapped siblings enabled WFs? The issue is that we have bunch of overlapped data WFs (even after L3 data unit) and people are stepping on each other, doubling the amounts every time someone loads their WF while other WF data load is in progress. Wanted to try event handler and message showing load is in progress. Any other ideas to avoid this are also appreciated. Thank you, Mikki40Views0likes1Comment