[OSM-dev] Importing OSM 0.5 with Osmosis
Brett Henderson
brett at bretth.com
Sun Mar 1 08:48:48 GMT 2009
Try searching google for "The total number of locks exceeds the lock
table size".
This link suggests increasing the mysql "innodb_buffer_pool_size" variable:
http://bugs.mysql.com/bug.php?id=15667
Frank Bielig wrote:
> Hallo,
>
> currently I'm running the OSM Server branch "rails_port_branches/api06".
> While importing a larger OSM 0.5 file with osmosis I get following
> error after several minutes:
>
> org.openstreetmap.osmosis.core.OsmosisRuntimeException: Unable to load
> current way nodes.
> at
> org.openstreetmap.osmosis.core.mysql.v0_6.MysqlWriter.complete(MysqlWriter.java:1044)
> at
> org.openstreetmap.osmosis.core.migrate.MigrateV05ToV06.complete(MigrateV05ToV06.java:82)
> at
> org.openstreetmap.osmosis.core.xml.v0_5.XmlReader.run(XmlReader.java:111)
> at java.lang.Thread.run(Thread.java:619)
> Caused by: java.sql.SQLException: The total number of locks exceeds the
> lock table size
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:1055)
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:956)
> at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:3491)
> at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:3423)
> at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:1936)
> at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2060)
> at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2542)
> at
> com.mysql.jdbc.PreparedStatement.executeInternal(PreparedStatement.java:1734)
> at
> com.mysql.jdbc.PreparedStatement.execute(PreparedStatement.java:995)
> at
> org.openstreetmap.osmosis.core.mysql.v0_6.MysqlWriter.complete(MysqlWriter.java:1041)
> ... 3 more
>
> For importing I use following command line with latest version of
> osmosis (SVN):
>
> osmosis --read-xml-0.5 file=file.osm.gz \
> --migrate \
> --write-mysql-0.6 host=<host> database=<db>
> user=<user> password=<password>
>
> Can anybody reproduce this error too? Is there any workaround?
>
>
> Thanks for any help!
> Frank
>
>
More information about the dev
mailing list