I have looked at several examples of cohorts and tried to regenerate some of them with Atlas.
Synpuf 1% (I assume it means 1% of the total synpuf) is tiny. Although the data source for the demos is miniscule and the inclusion criteria are very simple, this cohort builder took over a minute.
If I simply extrapolate the time to the full synpuf dataset (it is still TINY in comparison to real life EMR datasets) it would take over two hours. If I were to extrapolate this to TERABYTE sized datasets we would be looking at 1+ days.
Any question of “interactive” cohort building would be a complete fantasy.
Does anyone know what kind of hardware is being used to run the demo?
Assuming some people have installed the Atlas software in their environments, can you share your observations how long it takes to build cohorts and what computing resources you have to use for that?