I’m using the text backend because I’m running a largish model in a low-memory environment - an EC2 instance with only 3GB of RAM.
I’m not sure what the implications are of adding .save and .load methods to NDAarray are, but as long as there’s a way to run inference without storing the traces in memory, I’m happy.