How do you know if you are using the Roslyn compiler?
Hotfix 8.4.3 addresses, among other things, security vulnerabilities found in the third-party Roslyn Compiler DLL. How do you know if you are using the Roslyn compiler in your implementation of custom code and, therefore, at risk if you are? Is the Roslyn Compiler only used by the WinSCP libraries (and therefore its removal is WinSCP-related)? Or are there instances where Smart Integration Functions would be using the Roslyn compiler outside of the use of WinSCP?Solved17Views0likes1CommentWorkflow channels with scenario copies
Hi All, In our multi year planning process, most entities need to upload four sets of fixed costs (all loaded by the same team, but sent to that entity's finance team by different functional teams), in addition to all the driver-based costs we're calculating (mostly based on form inputs). For the four sets of fixed costs, we have four import profiles under the same parent workflow. The entities import their files and then run a DM sequence from a dashboard to run the validate/load for a range of years they can select. When they run this sequence, all import profiles (under the same parent) that have been imported are validated and loaded. We had not set up workflow channels on the different loads, and everything was working as expected. The different import profiles will use some of the same accounts, but different sets of cost centers. We did not see any collision of data when the same accounts were loaded, and if a balance that had previously been loaded in one import was removed from the import template and reimported, the data in that account would be cleared from the cube (even if that account was only used by one of the four imports) after the next validate/load. Now that we have started to copy cube data from one completed scenario (CY) to another blank scenario (CY_v2), so that some changes can be made to CY_v2, we were running into issues with our fixed cost loads in CY_v2 (balances not clearing, other values being overwritten). This makes sense, and we had not considered loading the fixed costs to copied scenarios. We would like to be able to only load one of the four imports (if say the G&A load is the only one with changes) in the CY_v2 scenario. We were hoping that workflow channels on the imports would solve the issue in the copied-to scenarios, but after completing that set up (changing accounts to nodatalock, setting up different data source members for each of the four imports, creating the channels, and applying the channels to the data source members), we are still seeing balances "stuck" in CY_v2 when previously loaded balances to a unique account are removed from the latest upload for one of the imports. We had also gone back and reloaded the "source" scenario with the channels applied, before copying the data and trying to load only a single template to CY_v2. Is there a way to use workflow channels so that all data loaded through that channel is cleared and replaced when a new import is loaded to a data source member where the channel is applied, or will accounts no longer in the latest import be left un-cleared? The documentation on the level 2 and level 3 data units seems to indicate that those accounts will not be cleared, but that does not match the behavior we see in the original loads of multiple sibling import profiles (described and bolded in my second paragraph). It also doesn't seem to match how a typical, iterative planning process would work. Any feedback would be appreciated. Thanks, James26Views0likes0CommentsError when processing of uploaded Excel file
Hi all, Could you please advise how to find error in Excel file which we try to load in OneStream on Import step of workflow. When I click Process button after uploading of file, I see the below message: Invalid URI: The hostname could not be parsed. Unable to execute Business Rule 'ADU_SolutionHelper'. A malformed URI was found in the document. Please provide a OpenSettings.RelationshipErrorRewriter to handle these errors while opening a package.23Views0likes2Comments"Cannot execute step because the prior workflow step is not completed"
We have Workflow A, and Workflow B. We get an error when the loadCube step in workflow B, when Workflow A is in Validation step. The error we get is "Cannot execute step because the prior workflow step is not completed. . (WP#WorkflowA.Import:S#Actual:T#2022M1)" The name the actual workflow has been changed to protect the innocent. Our question is: what is causing this issue and how do we fix it? It's not clear in the documentation what exactly triggers this error message. The parameters passed back imply that one cannot load to the same scenario and time if another workflow to that scenario and time is still open. WorkflowA and WorkflowB are loading to the same scenario and time, but to different accounts.4.8KViews1like5CommentsWorkflow Profile Hierarchies: Security Model | Application Security Roles v Workflow Security
Please share your practical advice to secure Workflow Profile Hierarchies per the two requirements below. Thank you. A. Workflow Profile Hierarchies: Admin v Builder/Business 0. 'Cube Root Workflow Profiles' | Admin and Business | Top Level Cube_suffix 1. 'Default/Child Workflow Profile Types' | Admin/Business 'Cube Root Workflow Profiles' 2. Admin 'Parent/Child Workflow Profile Types' | Admin/Business 'Cube Root Workflow Profiles' 3. Business 'Parent/Child Workflow Profile Types' | Business 'Cube Root Workflow Profiles B. Two Workflow Requirements: Admin v Builder/Business 1. Builder not allowed to see/create/edit: - Admin 'Cube Root Workflow Profiles' - 'Default/Child' and Admin 'Parent/Child Workflow Profile Types' | Admin 'Cube Root Workflow Profiles' - 'Default/Child Workflow Profile Types' | Business 'Cube Root Workflow Profiles' - Admin 'Parent/Child Workflow Profile Types' | Business 'Cube Root Workflow Profiles' 2. Admin/Builder allowed to see/create/edit: - Business 'Cube Root Workflow Profiles' -- Business 'Parent Workflow Profile Types': Review, Base Input, Parent Input -- Business 'Child Workflow Profile Types': Import, Forms, Journals C. Security Model: Application Security Roles + Workflow Security | Admin v Builder/Business 0. Security Groups: 'Admin Roles and Builder Roles' and Child Groups = 'Admin Roles' + 'Builder Roles' 1. Application Security Roles | Manage Workflow Profiles = 'Admin Roles and Builder Roles' 2. Workflow Security = 'Admin Roles' = Admin/Business 'Default Workflow Profile Types' 3.1. Workflow Security = 'Admin Roles' = Admin 'Cube Root Workflow Profiles' 3.2. Workflow Security = 'Admin Roles' = Admin 'Parent Workflow Profile Types' 4.1. Workflow Security = 'Builder Roles' = Business 'Cube Root Workflow Profiles' 4.2 Workflow Security = 'Builder Roles' = Business 'Parent Workflow Profile Types' D. Security Model Results: Admin v Builder/Business - 1 + 2 + 3 + 4 = Workflow Requirement 1 = Fail - 1 + 2 + 3 + 4 = Workflow Requirement 2 = Succeed Thank you, SMEs.230Views0likes1CommentWorkflow Import - Clearing cube data in v9
Hello, we recently upgraded to v9 and noticed that we can no longer clear WF imported cube data using the "Clear" option through the workflow import (snip below). Prior to the upgrade, we could clear the stage data, then run import/validate/load to clear the data that was imported to the cube. Now, this option only clears the data from stage and in order to clear the data loaded to the cube, we have to use a data management step. In our case, this is now difficult because the data loaded may be specific to workflow channels and also use multiple dimension criteria and not just Entity (used by DM step). Is anyone else having this issue? Are there any other methods to removing the imported data via csv file from the cube other than a DM step? If we try to run the old process in v9. Clear, then Retransform, Load to clear cube data. This error appears.133Views0likes6CommentsErrors creating New Task in Data Import Schedule Manager
I am getting several errors when setting up a New Task in the Data Import Schedule Manager. We are on OS version 9.01.17403, and I have installed the latest version of DSM, which is 8.4.0_SV100. I suspect this may be a version compatibility issue, so I am curious if anyone has been able to get this solution to work in a 9.01 application. I have already uninstalled and reinstalled the solution, which didn’t resolve the issues. Below are the two errors I am seeing: When choosing Global Scenario from the Scenario(s) drop down list, I get an immediate error “Error processing member. The item was not found. Member, 11111111.” The details state: Unable to execute Business Rule ‘DSM_Paramhelper’ where it appears to be trying to call the Global Scenario by OneStream.Client.Api.DashboardsAjaxServiceReference.DashboardsAjaxServiceClient.GetParameterDisplayInfosUsingDashboardNameCompressed(SessionInfo si, LoadDashboardInfo loadDashboardInfo, Boolean isForDashboardUIWithInteractiveComponents, Dictionary`2 custSubstVarsAlreadyResolved). If I pick a specific Scenario, I am getting a different error. It allows me to pick the Scenario and Time, but when save the Task, I get “The input string ‘’ was not in a correct format.”. The error details show it is an issue with the same Business Rule ‘DSM_SolutionHelper’ where Conversion from string “” to type ‘Double’ is not valid. The input string ‘’ was not in a correct format. OneStream.Client.Api.DashboardsAjaxServiceReference.DashboardsAjaxServiceClient.StartExecuteSelectionChangedServerTaskCompressed(SessionInfo si, Boolean isSystemLevel, Guid primaryDashboardID, Guid embeddedDashboardID, Guid componentID, PageInstanceInfo pageInstanceInfo, XFSelectionChangedServerTaskInfo serverTaskInfo) Any advice on how to correct these issues would be greatly appreciated.27Views1like0CommentsTransformation Rules - Possible Bug
I have an issue, since we migrated to V9, OS is adding on its own a space after transforming a member and this prevent us from loading the data to the member. (This is in the validation step when it fails) Data source has been checked, adjusted, and tested in many ways. The extra space doesn't come from here. Transformation rules have also been tested in many ways, and the space is also not here, we event tried one-to-one, masks, all ,and this is not the issue The error is only happening with 2 UD3 members US4 & CA4 both ending in 4, I created a fake one for testing "US41" and this one works fine. I have been checking with OS support, but they just want to repeat over and over the same testing we already did with them on a call To me, after checking all that I could, seems like a bug in the new V9.1, but if anyone has any ideas please let me know.The supplied Workflow Time is more than 3 periods after the last import period
We are currently in the process of upgrading to OFC PV840-SV112 and just applied the upgrade to our Customer Acceptance Testing Application where we have our imports and matching running daily. However, the amount of data imported is much less in this applicationas it is using a test database to import the data from and one of the matchsets has not imported any data for many months. The matching step for that matchset erred out a couple of days after the upgrade with the message "The supplied Workflow Time is more than 3 periods after the last import period". We discovered that if we loaded just one "dummy" row for the matchset that the error stopped happening. My questions are - has anyone else noticed this? And, if so, is there a setting somewhere to either turn this check on or off or increase the number of months that it looks back? Thanks for any information you can provide.16Views0likes0CommentsFTP File Load Best Practice
When I pull a flat file from an SFTP server, I have been parsing that file in VB .net into a Data Table and loading it that way. Is this the best way to do this, or is there a way to make the Import read the file from the filesystem once it has been downloaded? I feel like parsing a csv file using Split() is risky. Thanks, ScottSolved45Views0likes2Comments