在 MySQL 中实现快速插入大量数据的最佳方法是什么?

我用 C 语言编写了一个程序来解析大型 XML 文件,然后使用插入语句创建文件.其他一些进程会将文件摄取到 MySQL 数据库中.这些数据将用作索引服务,以便用户可以轻松找到文档.

I have written a program in C to parse large XML files and then create files with insert statements. Some other process would ingest the files into a MySQL database. This data will serve as a indexing service so that users can find documents easily.

我选择 InnoDB 是因为它具有行级锁定的能力.C 程序将在给定调用上生成 500 到 500 万条插入语句.

I have chosen InnoDB for the ability of row-level locking. The C program will be generating any where from 500 to 5 million insert statements on a given invocation.

尽快将所有这些数据输入数据库的最佳方法是什么?要注意的另一件事是数据库位于单独的服务器上.是否值得将文件移至该服务器以加快插入速度?

What is the best way to get all this data into the database as quickly as possible? The other thing to note is that the DB is on a separate server. Is it worth moving the files over to that server to speed up inserts?

此表不会真正更新,但行将被删除.

This table won't really be updated, but rows will be deleted.

推荐答案

  • 使用 mysqlimport 工具或 LOAD DATA INFILE 命令.
  • 暂时禁用不需要数据完整性的索引

相关文章