26 February 2015 7 2K Report

I have a good sense of how Stata and SAS differ in how they handle data files (with SAS reading only one observation into memory at a time and being able to handle multiple files at one, and Stata reading the entire file, and thus only handling one file). Some colleagues say they hate Stata for that limitation (particularly for data processing), but I've never found it that limiting (maybe because I do simple merging and matching mostly). Current Stata has automatic memory management, so no need to bother with that anymore. I'm almost 100% in Stata for all my analyses for all its features, pretty charts easily done, and flexibility. It always seems easier to program and learn, and I get so much more out of Stata Press books than I do out of SAS books. 

But I just got my largest data sets ever (1 to 5 M records and a couple dozen variables across data files). I'm a little worried that I'm going to get into Stata with it and get in a bind due to file size or a combination of that with data prep (e.g., aggregation, merging, transposing) and mathematical operations. I'm planning to run two-level multilevel logistic and multinomial models. 

Despite all my Googling, I can't find any SAS v. Stata benchmarks. I thought someone out here might have some (or enough experience to put me at ease). 

I'm running Stata 12 on a Win XP box (REALLY!) with 2.39 GHz proc and 3.25 GB RAM. My largest file is 320 MB so well within my RAM. 

Thanks in advance. 

More Matt Jans's questions See All
Similar questions and discussions