[OSM-talk] How to deal with data duplication in bulk upload
Maciek Kaliszewski
mkalkal at interia.pl
Thu Mar 20 12:10:53 GMT 2008
Hi,
Some informations about UMP http://ump.waw.pl/en/index.html ,
http://wiki.openstreetmap.org/index.php/UMP2OSM project. They collected
a lot of valuable data ,so it would be great to import it to OSM. They
constantly improve their data so keeping in sync OSM copy would be
nightmare . The only way is to persuade them to use OSM infrastructure
as main repository. It would require implementing tools to convert from
(garmin [POLYGON]/[POLYLINE] text files) to OSM (one time ) and
convering back from OSM to garmin (used when they want compile new map)
They keep their data in CVS and use such tools as diff ,patch to
maintain data ,so use OSM database would be step forward.
But main problem is how to deal with data duplication . There are areas
such as Warsaw well mapped in OSM . I think that UMP guys like to
use their data (at least for some time), So my idea is to convert all
data to OSM ,and in areas empty in OSM tag it with appropriate OSM tags.
In places where we suspect data duplication tag it with prefixed tag ,
say fixme_highway=secondary for example , so renderer will not be hurt.
But how to detect automatically if there is a data duplication . Find
two parallel ways spaced 10 - 50 m and similarly tagged .
Regards
Maciek Kaliszewski
----------------------------------------------------------------------
Dzwon tanio w Swieta!
kliknij >>> http://link.interia.pl/f1d62
More information about the talk
mailing list