-
Hi targets community, I'm interested in modeling a financial time series problem using targets. This poses a somewhat unique challenge: a substantial number of targets are downstream of a large data set which is updated incrementally, daily. Most of the computation downstream from this target can be done in parallel across the entire data set. Thus, most of the computation is not invalidated by one new day of data. If this download task is triggered monolithically then all downstream tasks will get invalidated daily and lots of recompilation is necessary. A simple fix is to split the download task into dynamic batches (say Unfortunately, this forces one download operation per What I'd like to do is to tap into one of two mechanisms to make this whole system efficient:
In either case, the need is to be able to access some historical or "meta" information in the target's command. Is there an existing mechanism to do something of this nature? Alternatively, do others with experience working on regularly-updating time series using targets have some best practices for this situation? Thanks, |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
It's actually possible to avoid most of that invalidation without getting too low-level. Sketch:
The full dataset will download every day, but the branches of |
Beta Was this translation helpful? Give feedback.
It's actually possible to avoid most of that invalidation without getting too low-level. Sketch:
The full dataset will download every day, but the branches of
analysis
will not rerun if the corresponding row groups…