[OSM-talk] How to deal with data duplication in bulk upload

Maciek Kaliszewski mkalkal at interia.pl
Thu Mar 20 12:10:53 GMT 2008


Hi,

Some informations about UMP http://ump.waw.pl/en/index.html , 
http://wiki.openstreetmap.org/index.php/UMP2OSM project. They collected 
a lot of valuable data ,so it would be great to import it to OSM. They 
constantly improve their data so keeping in sync OSM copy would be 
nightmare . The only way is to persuade them to use OSM infrastructure 
as main repository. It would require implementing tools to convert from 
(garmin [POLYGON]/[POLYLINE] text files) to OSM  (one time ) and 
convering back from OSM to garmin (used when they want compile new map)
They keep their data in CVS and use such tools as diff ,patch to 
maintain  data ,so use  OSM database would be step forward.

But main problem is how to deal with data duplication . There are areas 
such as Warsaw well mapped in OSM . I think that UMP guys  like  to  
use  their  data  (at least for some time), So my idea is to convert all 
data to OSM ,and in areas empty in OSM tag it with appropriate OSM tags.
In places where we suspect data duplication tag it with prefixed tag , 
say fixme_highway=secondary for example , so  renderer will not  be hurt.
But how to detect automatically if there is a data duplication  .  Find 
two parallel ways spaced 10 - 50 m and  similarly tagged .


Regards
Maciek Kaliszewski

----------------------------------------------------------------------
Dzwon tanio w Swieta!
kliknij >>>  http://link.interia.pl/f1d62 





More information about the talk mailing list