On Jul 25, 2012, at 4:58 PM, Fabrizio Perin wrote:
Sorry but the problem is not that there are no 64-bit
vms and images available. The problem is that the available 32-bit vm and image cannot be
pushed further 500MB. Even worst is that as far as I understood we are not even sure why
is like that.
Is that windows only? I run larger mac images.
For me a reasonable size for a Moose image containing
an average size Java Enterprise application is between 500MB and 1500MB. So a 32-bit
vm\image should be perfectly able to store the whole model and to have enough free space
For me the difference between a 500 MB model and a 2GB model is not really meaningful. It
a significant limit to the size of models I can handle. I try to avoid loading as much as
possible. A 588 MB
image starts in 3 seconds on my smallest machine, so that is fast enough.
Partial loading could be a solution in some cases but
we need tool support for that. I cannot invest 2 weeks every time I need to script a 10
minutes analysis trying to figure out how to partially load the information that I
"might" need. Without having a full model available the entire idea of
prototyping analysis behind Moose goes down the drain and so Moose itself lose a lot of
I think the whole point is to have all the data on the
system in analysis at hand. Either having 10GB model stored in an image or loading the
needed entities on demand it is not relevant as soon as it is transparent for the user and
the performances are not too bad.
I don't understand how performance can be good using a nosql or rdbms system. Gemstone
with enough ram,
or multiple pharo images with distributed processing, yes, but copying all that data
around sounds to me
like a non-starter.