On 01 Jan 2017, at 16:15, stephan stephan@stack.nl wrote: The more classical way to provide real-time data-analysis is using gemstone technology http://www.gemstone.com/customers/DISA http://www.gemstone.com/customers/DISA
I fully agree they should be be using that. However most companies don’t (want to) know Smalltalk anymore. For those RDF was an answer to their prayers and SPARQL the promise to be db-backend independent for eternity :-)
You could, and there might not be much added value in doing that instead of just creating a domain model directly in smalltalk.
That’s a main thing I have to research - which benefits do I get from using/implementing this with a matching meta model.
Moose expects a dynamic model, but one that is not so large. With a 64-bit image the size is less of a problem than the processing speed. You might want to distribute the processing over many images, and create different mappings of data sources to images to optimize your queries.
Distributed images … that’s something I didn’t think about yet. This might really solve some concerns I have with the backend.
By default log files are extremely redundant. Compressing them by using a decent domain model, introducing value objects to reduce data size can be useful.
True - if (only) log data would be the (only) problem. Having done this for another for another project (REST API forensics) we easily achieved 1:1000+ ratios in compression. Especially modelling the IPv6 addresses as Singletons and using a tree model for the paths/queries provided amazing compression. And once the paths were objects and not mere dumb strings the final verdict came fast.
Most databases work badly with this kind of problem. You might want to expicitly model the mapping of data to multiple images. You are basically creating a blackboard system
Never heard of if - but researching it definitely matched the idea I have in mind!!! This is a great pointer I need to investigate further. I hope to find more patterns there helping me to do the system “right” with less iterations. Although I did some research on the stuff I plan to build I found no matching paper on the subject whatsoever. It seems I was missing the right terms for concepts and ideas :-(
This probably needs a double registration. From a governance p.o.v. you need the actual process steps, and you also need a log of how you could have reached your conclusions with perfect hindsight http://www.ics.uci.edu/~taylor/classes/121/IEEE86_Parnas_Clement.pdf http://www.ics.uci.edu/~taylor/classes/121/IEEE86_Parnas_Clement.pdf
Another gem!
Moose supports creating your own tools very well. We did a data migration with Moose, and created 60+ glamour browsers for it, and daily visualisations of the process. We managed to keep things in a single image, though we ran out of memory several times. To me it sounds like you'd need to develop something to manage lots of images, starting new ones on demand. That fits well with our Pharo vision.
I reached the point where the decision for Moose/Pharo is already taken (positively!). I’m just not decided yet on the backend(s) side.
Thank you very much for your comments! Very helpful and highly appreciated!
CU,
Udo