DM Step Issue with Exporting Data with View Filters
Hi team, I'm trying to export data using a Data Management step with view filters. When I specify the view as V#Periodic,V#YTD, only V#Periodic data records is being extracted. If I reverse the order to V#YTD,V#Periodic, then only V#YTD data records get extracted. Has anyone faced this issue or knows how to correctly extract data with View dimension filters? Thanks in advance! 🙂Solved17Views0likes1CommentHow to Run Workspace Assembly from a Data Management Step In 6 Easy Steps
This was originally a question.It took a while to figure out, so to save someone else the hours of trial and error, here is the process I use in detail, start to finish, to run a Workspace-based Data Management step in a Workspace Assembly business rule. Here's a How-to step by step: List of abbreviations used below: DM: Data Management WS: Workspace MU: Maintenance Unit BR: Business Rule AS: Assembly 1. Workspace:Set the NameSpace Prefix - this is what gets filled in at runtime for __WSNameSpacePrefix 2. Assemblies: Note these properties 3. Create the Assembly Business Rule:Create an "Extender" business rule, this is done by right clicking files and selecting "Extensibility Rules" Add some test code, to verify correct configuration, such as BRApi.ErrorLog.LogMessage(si, "Hello from the new extender!"then code the way you would any other extender 4. Create The Data Management Group: Create a DataManagement Group in the same workspace that the Assembly that has your business rule is in 5. Create a Data Management Step - Create a DM step in the new DM Group, reference the rule using the pattern: Workspace.{WorkspacePrefix}.{AssemblyName}.{BusinessRuleName} In this case it would be: Workspace.Current.DMTestAS.DMTestStepExtender Note: we could also use 'DMTestWS' instead of 'Current'.We use Current because the DM Step is in the same WS as the Extender Rule. 7. Done! Test the DM Step with run button and check the error log(System Tab -> Logging -> Error Log),for your test code's output.309Views1like10CommentsBI Blend Table - What is the significance of sEntity?
I just created a BI Blend table and when I open the file using a data adapter, I see the data unit and aggregation fields all have an extract column prefixed with "s" (i.e., sEntity, sScenario, sAccount, sUD1, etc.). What do these columns represent?15Views0likes1CommentLock Out Sold Entities, Keep Their Impact!
Managing Sold Entities Without Losing Consolidation Results. Hi Community, We’re looking to hear how others handle this in OneStream: When an entity is sold in a specific period (e.g., 2025M5), the requirement is to: ✅ Retain its posted journal values in group consolidation ❌ Prevent any further use of the entity (journals, data loads, forms) The challenge we've encountered is that setting In Use = False — even after journals are posted and consolidation has been run — can cause OneStream to drop that entity’s data from the group during future reconsolidation. This results in the loss of valid P&L impact from the sold entity. Our workaround: Keep In Use = True for the sold period Use workflow locks and security restrictions to block future changes Only set In Use = False once everything is finalized and locked 💡 Feature Suggestion: It would be incredibly useful to have a system property or override flag that allows us to: Set In Use = False to fully disable the entity in the workflow Still retains its consolidated values during reconsolidation How are you managing this today? Your thoughts! Thanks,27Views0likes0CommentsTracking User Details and Timestamps in OneStream Dashboard Export
Hi , Good Day! I'm currently building a dashboard in OneStream that allows users to select all possible points of intersection for dimensions and their members. Once selected, they can click a button to export the data to a file share using a Data Management export sequence. This part is working well, and the dashboard is complete. Now, I want to enhance it by capturing user details (who entered the data) and the timestamp of when the data was entered for each point of intersection. I'm wondering if there's an existing OneStream Marketplace solution or feature that supports this kind of tracking and logging. Has anyone implemented something similar or come across a solution that could help with this? Thanks in advance!17Views0likes0CommentsCalculate values in different scenarios
Hi all, I tried to issue an api.Data:calculate with the following (fake) formula : A#ACCXXX:F:MYFLOW:S#MYSCEN = A#ACC001:F#CHI:S#ACT Of course OS complains because I'm reading from the current Scenario (ACT) but I want to write in a different Scenario that implies a different Data Unit (is it ?) While being in ACT scenario, if I would switch to DataBuffers to write in the destination Scenario (is it?) Could someone point me in the right direction ? I did this, but looks like it doesn't write a **bleep** ! Dim _srcDC As DataCell = api.Data.GetDataCell("A#ACC001:F#CHI:S#ACT") Dim _targetDB As New DataBuffer() Dim _dcNew As DataCell = New DataCell(_srcDC) _dcNew.SetScenario(api, "MYSCEN") ' This WON'T CHANGE <=================== _dcNew.SetAccount(api, "ACCXXX") ' This Change _dcNew.SetFlow(api, "MYFLOW") ' This Change _targetDB.SetCell(si, _dcNew) api.Data.SetDataBuffer(_targetDB, _destinationInfo) In the end, works fine... just doesn't change the Scenario... :-/ What am I doing wrong ? Thanks in advance for your help FabioG31Views0likes1CommentDynamic Cube Data Load from BiBlend Table
I am currently working on a Dynamic Cube with Data Binding Type (Share Data using Workspace Assembly). I was able to load Local currency data from BiBlend table and verify data using Cube View. Does Dynamic cube support Translation & Consolidation? I was unable to view translated data.Solved139Views0likes8CommentsWorkflow time format is weird
Hello ! I've noticed that the workflow time stored in the application is having a weird format, for instance : Workflow Time : 022008000 = 2022M5 Workflow Start Time : 2022003000 = 2022M1 Workflow End Time : 2022018000 = 2022M12 The only reason I see for this weird format is on the FX rate table (shown below), as it seems that 2022M1 is in the 4th column (hence 003 if we count column 0), 2022M5 in the 8th column and 2022M12 in the 18th column... --> Is it just a coincidence, or am I assuming correctly the date format ?Solved3.5KViews0likes5CommentsCube | Data Access | Data Cell Access Security/Slice Security by U1#Geography
Data Management Access Security Please share your expertise to set up a Cube's Data Cell Access Security / Slice Security to limit data read access as outlined below: Dimension: U1# - Geography Total_Geography = Child 1 + Child 2 = Total / Top North_America = Child 1 / NA International = Child 2 / Int'l Slice Security: Access Level - Read Data by Geography - in a CV and a QV Total_Geo_Data = Read Total data | Read NA data and Read Intl data NA_Data = Read NA data | no data access = International and Total Intl_Data = Read Int'l data | no data access = North America and Total TY for your practical advice - SMEs.Solved197Views0likes2Comments