I'm implementing a monitoring system using GPS vehicle and my BIG PROBLEMNB is this, I have 100k data (GPS data) per second parallel to process.
Data is sent from many GPS information (GPSID, latitude, longitude, time), 100k data, with the same format for the second, this information should I process it (process it in some way (?)) To show the route of each gps in a visual system.
My big question is:
As I can "take" this large amount of data being sent every second? and after that process them in the most efficient way to display travel information for each GPS-id
Perhaps the process of grouping? for each GPS-id? I have not really clear things, any additional ideas or functionality will be helpful.
I would like to know some ideas of how to process this much information that applications use?, Algorithms? divided into several computers (how?)? ...
the most likely solution, involving architecture and algorithm, im working with java.
I've been reading about hadoop and map/reduce and i don not know if I could serve.
I just found it: https://github.com/nathanmarz/storm
any idea is appreciated
Info ad: GPS information is created by a script, I must see what the best way to receive such information and probably if I can skip a block of information, since 1 second of lost data blocks of coordinates think not affect the display and gives me more time to process the data, are possibilities, obviously the less information lost is much better