mysql - 提高mysql导入速度

我有 22GB 的大型数据库。我曾经使用 gzip 格式的 mysqldump 命令进行备份。

当我提取 gz 文件时,它会生成 16.2GB

.sql 文件

当我尝试在本地服务器中导入数据库时​​,导入大约需要 48 小时。有没有办法提高导入过程的速度?

我也想知道是否需要进行任何硬件更改以提高性能。

当前系统配置

 Processor: 4th Gen i5
 RAM: 8GB

#update

my.cnf如下

#
# The MySQL database server configuration file.
#
# You can copy this to one of:
# - "/etc/mysql/my.cnf" to set global options,
# - "~/.my.cnf" to set user-specific options.
# 
# One can use all long options that the program supports.
# Run program with --help to get a list of available options and with
# --print-defaults to see which it would actually understand and use.
#
# For explanations see
# http://dev.mysql.com/doc/mysql/en/server-system-variables.html

# This will be passed to all mysql clients
# It has been reported that passwords should be enclosed with ticks/quotes
# escpecially if they contain "#" chars...
# Remember to edit /etc/mysql/debian.cnf when changing the socket location.
[client]
port        = 3306
socket      = /var/run/mysqld/mysqld.sock

# Here is entries for some specific programs
# The following values assume you have at least 32M ram

# This was formally known as [safe_mysqld]. Both versions are currently parsed.
[mysqld_safe]
socket      = /var/run/mysqld/mysqld.sock
nice        = 0

[mysqld]
#
# * Basic Settings
#
user        = mysql
pid-file    = /var/run/mysqld/mysqld.pid
socket      = /var/run/mysqld/mysqld.sock
port        = 3306
basedir     = /usr
datadir     = /var/lib/mysql
tmpdir      = /tmp
lc-messages-dir = /usr/share/mysql
skip-external-locking
#
# Instead of skip-networking the default is now to listen only on
# localhost which is more compatible and is not less secure.
bind-address        = 127.0.0.1
#
# * Fine Tuning
#
key_buffer      = 16M
max_allowed_packet  = 512M
thread_stack        = 192K
thread_cache_size       = 8
# This replaces the startup script and checks MyISAM tables if needed
# the first time they are touched
myisam-recover         = BACKUP
#max_connections        = 100
#table_cache            = 64
#thread_concurrency     = 10
#
# * Query Cache Configuration
#
query_cache_limit   = 4M
query_cache_size        = 512M
#
# * Logging and Replication
#
# Both location gets rotated by the cronjob.
# Be aware that this log type is a performance killer.
# As of 5.1 you can enable the log at runtime!
#general_log_file        = /var/log/mysql/mysql.log
#general_log             = 1
#
# Error log - should be very few entries.
#
log_error = /var/log/mysql/error.log
#
# Here you can see queries with especially long duration
#log_slow_queries   = /var/log/mysql/mysql-slow.log
#long_query_time = 2
#log-queries-not-using-indexes
#
# The following can be used as easy to replay backup logs or for replication.
# note: if you are setting up a replication slave, see README.Debian about
#       other settings you may need to change.
#server-id      = 1
#log_bin            = /var/log/mysql/mysql-bin.log
expire_logs_days    = 10
max_binlog_size         = 100M
#binlog_do_db       = include_database_name
#binlog_ignore_db   = include_database_name
#
# * InnoDB
#
# InnoDB is enabled by default with a 10MB datafile in /var/lib/mysql/.
# Read the manual for more InnoDB related options. There are many!
#
# * Security Features
#
# Read the manual, too, if you want chroot!
# chroot = /var/lib/mysql/
#
# For generating SSL certificates I recommend the OpenSSL GUI "tinyca".
#
# ssl-ca=/etc/mysql/cacert.pem
# ssl-cert=/etc/mysql/server-cert.pem
# ssl-key=/etc/mysql/server-key.pem



[mysqldump]
quick
quote-names
max_allowed_packet  = 512M

[mysql]
#no-auto-rehash # faster start of mysql but no tab completition

[isamchk]
key_buffer      = 512M

#
# * IMPORTANT: Additional settings that can override those from this file!
#   The files must end with '.cnf', otherwise they'll be ignored.
#
!includedir /etc/mysql/conf.d/

它正在上传 3 天,现在已导入 9.9 GB。数据库有 MyISAMInnoDB 表。我可以做些什么来提高导入性能?

我尝试使用 mysqldump 以 gz 格式分别导出每个表,并通过执行以下代码的 PHP 脚本导入每个表

$dir="./";
$files = scandir($dir, 1);
array_pop($files);
array_pop($files);
$tablecount=0;
foreach($files as $file){
    $tablecount++;
    echo $tablecount."     ";

    echo $file."\n";
    $command="gunzip < ".$file." | mysql -u root -pubuntu cms";

    echo exec($command);
}

最佳答案

以上述方式进行转储和恢复将意味着 MySQL 在导入数据时必须完全重建索​​引。它还必须每次解析数据。

如果你能以 MySQL 已经理解的格式复制数据文件,效率会高得多。这样做的一个好方法是使用 innobackupex来自 Percona

(开源并作为XtraBackup 的一部分分发,可从here 下载)。

这将拍摄 MyISAM 表的快照,对于 InnoDB 表,它将复制底层文件,然后针对它们重放事务日志以确保一致的状态。它可以从实时服务器上执行此操作而无需停机(我不知道这是否是您的要求?)

我建议您阅读文档,但要以最简单的形式进行备份:

$ innobackupex --user=DBUSER --password=DBUSERPASS /path/to/BACKUP-DIR/
$ innobackupex --apply-log /path/to/BACKUP-DIR/

如果数据在同一台机器上,那么 innobackupex 甚至还有一个简单的恢复命令:

$ innobackupex --copy-back /path/to/BACKUP-DIR

还有更多选项和不同的实际备份方式,因此我真的建议您在开始之前仔细阅读文档。

作为速度引用,我们的慢速测试服务器大约 600 IOPS,使用这种方法可以在大约 4 小时内恢复 500 GB 的备份。

最后:您提到了加快导入速度的方法。这主要取决于瓶颈是什么。通常,导入操作是受 I/O 限制的(您可以通过检查 io 等待来测试这一点),加快速度的方法是使用更快的磁盘吞吐量 - 或者更快的磁盘本身,或者同时使用更多的磁盘。

https://stackoverflow.com/questions/29643714/

相关文章:

linux - 我想为 Linux 内核做贡献

python - 如何用python替换像文本一样的sed?

python - 在亚马逊 ec2 linux 微型实例上的 virtualenv 中安装 scip

linux - 如何使用 Ansible 等待服务器重启?

linux - 如何计算 Bash 变量中的字符数

ruby - 如何在 linux (ubuntu) 上更新 ruby​​?

linux - 如何在 Linux 上从包含多个分区的镜像文件中挂载一个分区?

linux - 在 Bash 中将文本文件作为命令运行

linux - ctrl-x 在终端中使用时会发送哪个信号?

linux - 如何用逗号而不是空格分割列表