Onestream Certification : Lead architect or Core application architect ?
Hello everyone, I recently attempted to register for a OneStream certification exam and noticed that both the Lead Architect and Core Application Architect certifications are available. Could you please clarify which one is required for my consulting firm to qualify for a higher-level partnership with OneStream? Additionally, I would appreciate any insights into the main differences between these two certifications, as well as the passing score (percentage of correct answers) for each. Finally, if you have already taken one of these exams, I would be grateful if you could share your experience. Thank you in advance!Solved75Views0likes2CommentsCan constraints Slow Things Down in OneStream?
Constraints in OneStream are filters applied to metadata dimensions such as Flow, Accounts, UD1, or specific cash flow lines to narrow the scope of calculations, data retrieval, or Cube View rendering. They’re powerful tools, but when applied too broadly or dynamically, can they degrade performance? Larger evaluation footprint A Flow or Account constraint forces OneStream to check every intersection in that dimension. In dense cubes, this can mean millions of evaluations. 1. Dynamic filters are expensive. Static lists (e.g., A#1000, A#2000, Flow#Operating) resolve quickly. Dynamic logic FilterDim, Boolean conditions, or formula driven sets require runtime computation. Why? 2. Stacked or nested constraints multiply the cost. Does layering constraints across Flows, Accounts, and UDs in OneStream force the system to resolve each filter before executing calculations? 3. Reduced parallelism. When constraints slice across cash flow lines or UD dimensions, the engine can’t fully parallelize tasks. Does more constraint mean fewer threads working freely? Do you agree? Why? Cheers38Views0likes0CommentsArchive applications
Hi all, How do other customers go about archiving applications? We've been live since 2017 and every year after filing our SEC reporting, we copy the production application and rename it with the year that's being archived. This is helpful as in the past we've move products in segments as well as various other master data changes affecting reporting. The issue is that now we're nearing our storage capacity as the size of our application has grown over the years as we add in more data and features (35GB for the 2018 archive while our current 2021 production app is at 155 GB) and clearly this problem will continue every year. Any suggestions on a more efficient and cost-effective strategy? Thanks!2KViews0likes3CommentsDatabase space control policies
Hi! We are trying to control the size increase in our production application. We know the growth comes from: Stage tables that allow users to drill down and back onto source data. Data records Audit tables (they can grow a lots if they are recording continuous changes to rules, data for example) Logs which will impact Framework database growth (task activity and error logs, will vary depending on the detail login set in BR, DM and on the system configuration as well as your retention policy) Load transactional data, which will increate your data units and data records unnececarily, which BIBlend should be used for this type of data load instead of loading it to the finance cube. We have analyzed it and we realized the StageAttributeTable and other stage tables has a significant size. Has ayone performed databse maintenance by deleting data from the stage tables? Is it possible to do that while keeping the cube data intact and workflows completed? Any ideas or best practices regarding database maintenance would be very welcome. Regards, Xavi50Views0likes2CommentsHow do I change a parameter value in a dashboard?
Hi everyone, I have a business rules that is triggered through a dashboard. This Business rule is looking at a True/False parameter. How do I add the parameter to the dashboard and with a dropdown list showing the valid values (true/false) that can be used to save the value to the parameter? Thanks in advance,73Views0likes4CommentsCube | Data Access | Data Cell Access Security/Slice Security by U1#Geography
Data Management Access Security Please share your expertise to set up a Cube's Data Cell Access Security / Slice Security to limit data read access as outlined below: Dimension: U1# - Geography Total_Geography = Child 1 + Child 2 = Total / Top North_America = Child 1 / NA International = Child 2 / Int'l Slice Security: Access Level - Read Data by Geography - in a CV and a QV Total_Geo_Data = Read Total data | Read NA data and Read Intl data NA_Data = Read NA data | no data access = International and Total Intl_Data = Read Int'l data | no data access = North America and Total TY for your practical advice - SMEs.Solved300Views1like3CommentsManagement vs Stat reporting - best practice approach?
Hello, We want to review both our Management and Stat reporting on a monthly basis. The 2 approaches we are considering using are: employ an Audit / Nature UD to build the separate views by adjustments layered over the local GAAP import. or use a separate Scenario for each view with the Local GAAP data being fed into each scenario. One of the main differences between the two views would be pre-acquisition TBs for newly acquired companies being included in the management view for comparative purposes. Would be Interested to hear anyone's feedback or suggestions - thanks!45Views0likes1CommentInvalid Cube Name Error During Data Load
My colleagues in OneStream, have you seen this error before? Since this is the first time I’ve come across it in a standard load scenario with everything else validating, I thought I’d share here to compare notes. Any insights on what could be “behind” this error in practice would be really helpful for building a knowledge base around less common OneStream load errors. While running a data load in OneStream, I encountered an "Invalid Cube Name" error. What made this surprising is that all the usual suspects checked out fine: Workflow setup was validated. Data sources were reviewed. Transformation rules were checked. Rules on members who call cubes were also confirmed. Cube mappings looked correct. Despite all of these checks, the error still occurred. This message is unusual because it doesn’t typically appear if the metadata, cube structure, or mappings are aligned. In most cases, it points to something deeper in how OneStream is resolving the cube context for the workflow profile or the entity being processed. Have you seen this error before on OneStream? ( Invalid Cube Name Error During Data Load) Cheers85Views0likes1Comment