Luis Chaves Rodriguez
03/20/2025, 6:05 PMHall
03/20/2025, 6:05 PMdatajoely
03/20/2025, 6:11 PMLuis Chaves Rodriguez
03/20/2025, 6:13 PMdatajoely
03/20/2025, 6:16 PMNok Lam Chan
03/20/2025, 8:50 PMNok Lam Chan
03/20/2025, 8:52 PMLuis Chaves Rodriguez
03/22/2025, 6:30 AMOr do you mean you need conditional logic to create pipeline? That should also be doable.In some instances, different clients might need slightly tweaked configurations
Luis Chaves Rodriguez
03/22/2025, 6:32 AMWhat I’m really not clear about is, if we don’t manage to abstract away the logic to build one pipeline to cover all the scenarios would we end up with let’s say 4 pipelines (preprocess, fe, train, evaluations) x 8 problems scenarios (32 atomic pipelines and 8 “problem” pipelines) - if so how is it recommended to managed that? At that point is each of these in a separate project/repo? Is Alloy designed for this sort of scenario?
Also, let’s say our end to end pipelines only ever really runs once or once in a while to train an initial model on historical data and on data drift events or when a new features are added and we want the changes to apply to all of our clients. But then we have some pipelines that run more often such as getting predictions after loading a model that’s already fit or re fitting to latest data and getting predictions. Should this live in the same repo? Should they be all part of one big pipeline and some pipelines are skipped conditionally?