Execution Timeout Expired
Error Message: Execution Timeout Expired. The timeout period elapsed prior to completion of the operation or the server is not responding. Operation cancelled by user. The statement has been terminated. (SQL Error Number -2146232060, -2, 0, 3621). The wait operation timed out This happened after running Load and Transform for one of my import workflows. Any idea what's causing this? It stops at the PostFileArchives stage:23Views0likes1CommentFTP File Load Best Practice
When I pull a flat file from an SFTP server, I have been parsing that file in VB .net into a Data Table and loading it that way. Is this the best way to do this, or is there a way to make the Import read the file from the filesystem once it has been downloaded? I feel like parsing a csv file using Split() is risky. Thanks, ScottData Cell is read-only because Parent Workflow Profile has no active Input Profiles
Hi all, I have an Import, Validate, Load workflow that hasn't been used for a while. I'm trying to load in data and I'm getting invalid intersections at the validation stage with the following error message: "Data Cell is read-only because Parent Workflow Profile 'Workflow Name' has no active Input Profiles for the Workflow Channels that are assigned to the Data Cell's Account and/or UD member. Entity = Entity, Account=Volume, Origin=Import" Note I've use italics instead of using our entity names and workflow names. We didn't have this issue before so I'm not sure what has changed. The parent workflow profile stated in the validation message isn't actually the parent workflow profile so I'm totally confused. There is another post on this but I didn't understand the resolution. Could somebody please tell me what I would need to do to fix this? Thanks!Solved4.3KViews0likes5CommentsError when accessing the Workflow Import step
Hi Everyone, I was recently working within a Demo Environment where I suddenly started getting error message when trying to access the import step in the Load and Transform (Import) step in the Import Workflow. The Error above pops up every time anyone tries to access the Load and Transform (Import) step. Have anyone seen this issue before and knows how to fix it? NOTE: This issue is the same across all applications in this environment. Thanks, HamzaSolved278Views0likes4CommentsBusiness Rule | One-to-One Transformation - E# Parent to its first E# Base
We need your help to Import flat files: 1. Number of Rows: > 100k 2. Organized under E# Parent Members: a few thousands 3. E# is extended We've been fortunate to be advised to build a BR to generate a one-to-one Transformation Rule file. - E# Parent to its first E# Base Is it possible for you SMEs to be generous enough to share with us/Community: - a sample BR to look up its first E# Base Member of an E# Parent Member - under extended E# - how to call that BR in an Import Workflow Thank you, OS SMEs.94Views0likes2Commentsimporting a file with variable columns
We have been tasked with importing a file that has a large number of variable columns. For the sake of easy explanation, let's say the first five columns are standard (time, entity, ud1, ud2, ud3) but the file could have from 50 to 150 additional columns, one for each account. If there is no data for the account, there is no column for it. New accounts could appear in the future without warning. No, we don't have the ability to change the format of the report. (Oh, how I wish.) I have thought up several ways of making this work but each is fraught with its own type of peril. Create a new, custom table dynamically to stage the data. Parse the column names from the file. Use the column name list to run a new SQL query to unpivot. Parse the file in-memory to manually unpivot by parsing each data column and adding rows to a datatable, then returing the full data table. Maintain a list of the columns we care about the most, parse the file in advance and save the column name/position maps to parameters/a lookup table. Use up every possible attribute/value field in a data source to stage to BI Blend and try to unpivot from there. Hope they never need more "important" columns than OS can handle. (This is similar to option 1 but we're not stuck dropping/creating a custom table ourselves and we have more consistent column names.) Write a manual file-parser that creates a new, sane text file and then imports that instead. (Seems wasteful. If I can get it this far, I can probably just do it in-memory, ie, option 2.) Some other, better idea that I haven't thought of yet.Solved55Views0likes2CommentsHow to convert a raw Excel file into multiple upload-ready OneStream journals. (3 quick steps)
Use case: You generated several journal lines in Excel. You saved it as a CSV file, and the upload to OneStream failed. Here’s how to clean it up and get it posted. 1) Paste your raw lines into Excel: Paste your raw journal lines into column A as above. Go to Data → Text to Columns → Delimited → Comma → Finish. You should now see the first column as H and the next two as D, D, with each field in its column. 2) Clean the values (common blockers) ConsName must be a valid Consolidation member in your app (e.g., Entity or Local). Leave ParentName blank unless you’re 100% sure of the direct parent. Don’t type literal None unless it’s a real member; leave IC/UD blank or use your member when required. Make sure Scenario → Cube mapping matches your CubeName. If JournalBalanceType=Balanced, totals of Debits = Credits per journal. 2) Save as CSV (UTF-8) Keep only the upload rows on the active sheet (no notes above/beside them). File → Save As → CSV UTF-8 (Comma delimited) (.csv). Accept “only the active sheet will be saved.” 3) Import & post in OneStream OnePlace → Journals → Import (or your Workflow step). Choose File Type = CSV, select your file, and Import. Open the journal, Validate/Balance, then Post. Run Consolidation and verify in a Cube View. Hope this helps some people within the OneStream Universe 😊.How to copy a scenario from one application to another?
Hello, I recently recovered a scenario from a restored application copy from a point in time. I now would like to extract that scenario from the application copy and put that in our working application. Does anyone have any idea on how this would be done? Thank you, Jeremy MorganSolved1.7KViews0likes7Comments