FTP File Load Best Practice
When I pull a flat file from an SFTP server, I have been parsing that file in VB .net into a Data Table and loading it that way. Is this the best way to do this, or is there a way to make the Import read the file from the filesystem once it has been downloaded? I feel like parsing a csv file using Split() is risky. Thanks, ScottIdentifying API Parameters for OneStream Audit Log Export via Postman
Hello all, I joined a company where they have implemented OneStream solution. I have access to the platform, and still exploring the way it was implemented. I am currently working on integrating OneStream with an external platform and need to export audit logs using the product's API. While setting up the request in Postman, i've encountered difficulty identifying the correct values for the following parameters: ApplicationName AdapterName Despite reviewing the available documentation, these tags remain unclear. Has anyone successfully queried audit logs via the OneStream API and can share insights on where these parameters are defined or how to retrieve them? Any guidance or examples would be greatly appreciated. Thank you in advance!27Views0likes1CommentLoading Amount and Annotations together
Hi, not sure if anyone came across with this one but thought it was interesting to share If you have a file which has in the same line amount and annotation you can set View to YTD (or periodic), then in the text value complex expression use this api prop: api.Parser.TextValueRowViewMember = "Annotation" That will make the parser to automatically generate the line(s) with the annotation if text value have a value. HTHSolved1.7KViews8likes4Commentsimporting a file with variable columns
We have been tasked with importing a file that has a large number of variable columns. For the sake of easy explanation, let's say the first five columns are standard (time, entity, ud1, ud2, ud3) but the file could have from 50 to 150 additional columns, one for each account. If there is no data for the account, there is no column for it. New accounts could appear in the future without warning. No, we don't have the ability to change the format of the report. (Oh, how I wish.) I have thought up several ways of making this work but each is fraught with its own type of peril. Create a new, custom table dynamically to stage the data. Parse the column names from the file. Use the column name list to run a new SQL query to unpivot. Parse the file in-memory to manually unpivot by parsing each data column and adding rows to a datatable, then returing the full data table. Maintain a list of the columns we care about the most, parse the file in advance and save the column name/position maps to parameters/a lookup table. Use up every possible attribute/value field in a data source to stage to BI Blend and try to unpivot from there. Hope they never need more "important" columns than OS can handle. (This is similar to option 1 but we're not stuck dropping/creating a custom table ourselves and we have more consistent column names.) Write a manual file-parser that creates a new, sane text file and then imports that instead. (Seems wasteful. If I can get it this far, I can probably just do it in-memory, ie, option 2.) Some other, better idea that I haven't thought of yet.Solved53Views0likes2CommentsHow to copy a scenario from one application to another?
Hello, I recently recovered a scenario from a restored application copy from a point in time. I now would like to extract that scenario from the application copy and put that in our working application. Does anyone have any idea on how this would be done? Thank you, Jeremy MorganSolved1.7KViews0likes7CommentsCan ERP transactional data be loaded into OS?
Hi, my company's OS data integration with the ERP system has been set-up to load monthly TB's into OS and for transactional tables to sit outside of OS but with the ability for OS to query and drill-back to view transactions. The transactional drill-back function is limited in that a drill-down first needs to be performed to get to the detailed dimensional intersections at the lowest UD level before a drill-back to view transactions (at that precise lowest level set of dimensional intersections) can be run. It is therefore easier to use a stand-alone BI viewer outside of OS to view transactions. Am wondering - is it possible to load ERP transactional tables into OS (this may contain millions of rows spanning multiple years) and for OS to then display transactional level detail across higher-level roll-ups (e.g. viewing transactions across a cost centre parent rather than only at an individual cost centre level), or even display transactional detail in quick views, essentially making OS have similar functionality to a BI viewer?119Views0likes8CommentsLoading Annotations Only via a Workflow
Is it possible to create a data source to load annotations only from a flat file. I am tagging the annotations in the flat file with the View dimension. It seems like the data source always wants an "Amount". Is this case I have no amount so when I Ioad the data file, I get an error that "No Valid Data Keys in File". It seems that this should be possible. Maybe I should do this via Excel loads.Solved3.7KViews1like10CommentsAudit History for Forms - CubeView drill down (v8.4)
I have drilled down to the bottom of both Origin and UD6 dimension which shows the data was loaded by corp accounting team. However, i cannot view Audit history for Forms - it continues to show as grey'd out. Does anyone know how to enable the audit history? thank you.43Views0likes4Comments