一、描述
使用docker创建fastdfs,使用的镜像是season/fastdfs,然后对镜像进行了一些调整,并且改名为yxq18509376997/fastdfs:v1.1,当前我使用的是yxq18509376997/fastdfs:v1.1,其实使用season/fastdfs也是可以的。
二、准备
1> 下载镜像
docker pull season/fastdfs
或
docker pull yxq18509376997/fastdfs
我这里使用的是yxq18509376997/fastdfs
2> 准备目录
创建如下目录:
/home/docker/fastdfs/storage23000
/home/docker/fastdfs/storage24000
/home/docker/fastdfs/tracker22122
/home/docker/fastdfs/tracker22123
/home/docker/fastdfs/tracker22124
3> 准备tracker配置文件
配置文件总共是3个,但是修改的地方只有一个就是port,分别是22122,22123,22124;
分别放在如下目录:
/home/docker/fastdfs/tracker22122
/home/docker/fastdfs/tracker22123
/home/docker/fastdfs/tracker22124
# 需要自行配置 /fdfs_conf/storage
# is this config file disabled
# false for enabled
# true for disabled
# 该配置文件是否不生效,false 生效;true 不生效
disabled=false
# bind an address of this host
# empty for bind all addresses of this host
# 是否绑定IP,后面为绑定的IP地址 (常用于服务器有多个IP但只希望一个IP提供服务)。如果不填则表示所有的
bind_addr=
# the tracker server port
# 跟踪器服务端口
port=22122
# connect timeout in seconds
# default value is 30s
# 连接超时时间,针对socket套接字函数connect
connect_timeout=30
# network timeout in seconds
# default value is 30s
# tracker server的网络超时,单位为秒。发送或接收数据时,如果在超时时间后还不能发送或接收数据,则本次网络通信失败。
network_timeout=60
# the base path to store data and log files
# 目录地址(根目录必须存在,子目录会自动创建)
# 附目录说明:
# tracker server目录及文件结构:
# ${base_path}
# |__data
# | |__storage_groups.dat:存储分组信息
# | |__storage_servers.dat:存储服务器列表
# |__logs
# |__trackerd.log:tracker server日志文件
base_path=/fastdfs/tracker
# max concurrent connections this server supported
# 系统提供服务时的最大连接数。对于V1.x,因一个连接由一个线程服务,也就是工作线程数。
# 对于V2.x,最大连接数和工作线程数没有任何关系
max_connections=256
# accept thread count
# default value is 1
# since V4.07
# 允许线程数
accept_threads=1
# work thread count, should <= max_connections
# default value is 4
# since V2.00
# V2.0引入的这个参数,工作线程数,通常设置为CPU数
work_threads=4
# the method of selecting group to upload files
# 0: round robin
# 1: specify group
# 2: load balance, select the max free space group to upload file
# 上传组(卷) 的方式 0:轮询方式 1: 指定组 2: 平衡负载(选择最大剩余空间的组(卷)上传)
# 这里如果在应用层指定了上传到一个固定组,那么这个参数被绕过
store_lookup=2
# which group to upload file
# when store_lookup set to 1, must set store_group to the group name
# 当上一个参数设定为1 时 (store_lookup=1,即指定组名时),必须设置本参数为系统中存在的一个组名。如果选择其他的上传方式,这个参数就没有效了。
store_group=group1
# which storage server to upload file
# 0: round robin (default)
# 1: the first server order by ip address
# 2: the first server order by priority (the minimal)
# 选择哪个storage server 进行上传操作(一个文件被上传后,这个storage server就相当于这个文件的storage server源,会对同组的storage server推送这个文件达到同步效果)
# 0: 轮询方式
# 1: 根据ip 地址进行排序选择第一个服务器(IP地址最小者)
# 2: 根据优先级进行排序(上传优先级由storage server来设置,参数名为upload_priority)
store_server=0
# which path(means disk or mount point) of the storage server to upload file
# 0: round robin
# 2: load balance, select the max free space path to upload file
# 选择storage server 中的哪个目录进行上传。storage server可以有多个存放文件的base path(可以理解为多个磁盘)。
# 0: 轮流方式,多个目录依次存放文件
# 2: 选择剩余空间最大的目录存放文件(注意:剩余磁盘空间是动态的,因此存储到的目录或磁盘可能也是变化的)
store_path=0
# which storage server to download file
# 0: round robin (default)
# 1: the source storage server which the current file uploaded to
# 选择哪个 storage server 作为下载服务器
# 0: 轮询方式,可以下载当前文件的任一storage server
# 1: 哪个为源storage server 就用哪一个 (前面说过了这个storage server源 是怎样产生的) 就是之前上传到哪个storage server服务器就是哪个了
download_server=0
# reserved storage space for system or other applications.
# if the free(available) space of any stoarge server in
# a group <= reserved_storage_space,
# no file can be uploaded to this group.
# bytes unit can be one of follows:
### G or g for gigabyte(GB)
### M or m for megabyte(MB)
### K or k for kilobyte(KB)
### no unit for byte(B)
### XX.XX% as ratio such as reserved_storage_space = 10%
# storage server 上保留的空间,保证系统或其他应用需求空间。可以用绝对值或者百分比(V4开始支持百分比方式)。
#(指出 如果同组的服务器的硬盘大小一样,以最小的为准,也就是只要同组中有一台服务器达到这个标准了,这个标准就生效,原因就是因为他们进行备份)
reserved_storage_space = 10%
#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
# 选择日志级别
log_level=info
#unix group name to run this program,
#not set (empty) means run by the group of current user
# 操作系统运行FastDFS的用户组 (不填 就是当前用户组,哪个启动进程就是哪个)
run_by_group=
#unix username to run this program,
#not set (empty) means run by current user
# 操作系统运行FastDFS的用户 (不填 就是当前用户,哪个启动进程就是哪个)
run_by_user=
# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" means match all ip addresses, can use range like this: 10.0.1.[1-15,20] or
# host[01-08,20-25].domain.com, for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
# 可以连接到此 tracker server 的ip范围(对所有类型的连接都有影响,包括客户端,storage server)
allow_hosts=*
# sync log buff to disk every interval seconds
# default value is 10 seconds
# 同步或刷新日志信息到硬盘的时间间隔,单位为秒
# 注意:tracker server 的日志不是时时写硬盘的,而是先写内存。
sync_log_buff_interval = 10
# check storage server alive interval seconds
# 检测 storage server 存活的时间隔,单位为秒。
# storage server定期向tracker server 发心跳,如果tracker server在一个check_active_interval内还没有收到storage server的一次心跳,
# 那边将认为该storage server已经下线。所以本参数值必须大于storage server配置的心跳时间间隔。通常配置为storage server心跳时间间隔的2倍或3倍。
check_active_interval = 120
# thread stack size, should >= 64KB
# default value is 64KB
# 线程栈的大小。FastDFS server端采用了线程方式。更正一下,tracker server线程栈不应小于64KB,不是512KB。
# 线程栈越大,一个线程占用的系统资源就越多。如果要启动更多的线程(V1.x对应的参数为max_connections,V2.0为work_threads),可以适当降低本参数值。
thread_stack_size = 64KB
# auto adjust when the ip address of the storage server changed
# default value is true
# 这个参数控制当storage server IP地址改变时,集群是否自动调整。注:只有在storage server进程重启时才完成自动调整。
storage_ip_changed_auto_adjust = true
# storage sync file max delay seconds
# default value is 86400 seconds (one day)
# since V2.00
# V2.0引入的参数。存储服务器之间同步文件的最大延迟时间,缺省为1天。根据实际情况进行调整
# 注:本参数并不影响文件同步过程。本参数仅在下载文件时,判断文件是否已经被同步完成的一个阀值(经验值)
storage_sync_file_max_delay = 86400
# the max time of storage sync a file
# default value is 300 seconds
# since V2.00
# V2.0引入的参数。存储服务器同步一个文件需要消耗的最大时间,缺省为300s,即5分钟。
# 注:本参数并不影响文件同步过程。本参数仅在下载文件时,作为判断当前文件是否被同步完成的一个阀值(经验值)
storage_sync_file_max_time = 300
# if use a trunk file to store several small files
# default value is false
# since V3.00
# V3.0引入的参数。是否使用小文件合并存储特性,缺省是关闭的。
use_trunk_file = false
# the min slot size, should <= 4KB
# default value is 256 bytes
# since V3.00
# V3.0引入的参数。
# trunk file分配的最小字节数。比如文件只有16个字节,系统也会分配slot_min_size个字节。
slot_min_size = 256
# the max slot size, should > slot_min_size
# store the upload file to trunk file when it's size <= this value
# default value is 16MB
# since V3.00
# V3.0引入的参数。
# 只有文件大小<=这个参数值的文件,才会合并存储。如果一个文件的大小大于这个参数值,将直接保存到一个文件中(即不采用合并存储方式)。
slot_max_size = 16MB
# the trunk file size, should >= 4MB
# default value is 64MB
# since V3.00
# V3.0引入的参数。
# 合并存储的trunk file大小,至少4MB,缺省值是64MB。不建议设置得过大。
trunk_file_size = 64MB
# if create trunk file advancely
# default value is false
# since V3.06
# 是否提前创建trunk file。只有当这个参数为true,下面3个以trunk_create_file_打头的参数才有效
trunk_create_file_advance = false
# the time base to create trunk file
# the time format: HH:MM
# default value is 02:00
# since V3.06
# 提前创建trunk file的起始时间点(基准时间),02:00表示第一次创建的时间点是凌晨2点。
trunk_create_file_time_base = 02:00
# the interval of create trunk file, unit: second
# default value is 38400 (one day)
# since V3.06
# 创建trunk file的时间间隔,单位为秒。如果每天只提前创建一次,则设置为86400
trunk_create_file_interval = 86400
# the threshold to create trunk file
# when the free trunk file size less than the threshold, will create
# the trunk files
# default value is 0
# since V3.06
# 提前创建trunk file时,需要达到的空闲trunk大小
# 比如本参数为20G,而当前空闲trunk为4GB,那么只需要创建16GB的trunk file即可。
trunk_create_file_space_threshold = 20G
# if check trunk space occupying when loading trunk free spaces
# the occupied spaces will be ignored
# default value is false
# since V3.09
# NOTICE: set this parameter to true will slow the loading of trunk spaces
# when startup. you should set this parameter to true when neccessary.
#trunk初始化时,是否检查可用空间是否被占用
trunk_init_check_occupying = false
# if ignore storage_trunk.dat, reload from trunk binlog
# default value is false
# since V3.10
# set to true once for version upgrade when your version less than V3.10
# 是否无条件从trunk binlog中加载trunk可用空间信息
# FastDFS缺省是从快照文件storage_trunk.dat中加载trunk可用空间,
# 该文件的第一行记录的是trunk binlog的offset,然后从binlog的offset开始加载
trunk_init_reload_from_binlog = false
# if use storage ID instead of IP address
# default value is false
# since V4.00
# 是否使用server ID作为storage server标识
use_storage_id = false
# specify storage ids filename, can use relative or absolute path
# since V4.00
# use_storage_id 设置为true,才需要设置本参数
# 在文件中设置组名、server ID和对应的IP地址,参见源码目录下的配置示例:conf/storage_ids.conf
storage_ids_filename = storage_ids.conf
# id type of the storage server in the filename, values are:
## ip: the ip address of the storage server
## id: the server id of the storage server
# this paramter is valid only when use_storage_id set to true
# default value is ip
# since V4.03
id_type_in_filename = ip
# if store slave file use symbol link
# default value is false
# since V4.01
# 存储从文件是否采用symbol link(符号链接)方式
# 如果设置为true,一个从文件将占用两个文件:原始文件及指向它的符号链接。
store_slave_file_use_link = false
# if rotate the error log every day
# default value is false
# since V4.02
# 是否定期轮转error log,目前仅支持一天轮转一次
rotate_error_log = false
# rotate error log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.02
# error log定期轮转的时间点,只有当rotate_error_log设置为true时有效
error_log_rotate_time=00:00
# rotate error log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
# error log按大小轮转
# 设置为0表示不按文件大小轮转,否则当error log达到该大小,就会轮转到新文件中
rotate_error_log_size = 0
# if use connection pool
# default value is false
# since V4.05
# 是否使用连接池
use_connection_pool = false
# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
# 连接池的最大空闲时间 单位秒
connection_pool_max_idle_time = 3600
# HTTP port on this tracker server
# 以下是关于http的设置了 默认编译是不生效的 要求更改 #WITH_HTTPD=1 将 注释#去掉 再编译
# 关于http的应用 说实话 不是很了解 没有见到
http.server_port=8080
# check storage HTTP server alive interval seconds
# <= 0 for never check
# default value is 30
http.check_alive_interval=30
# check storage HTTP server alive type, values are:
# tcp : connect to the storge server with HTTP port only,
# do not request and get response
# http: storage check alive url must return http status 200
# default value is tcp
http.check_alive_type=tcp
# check storage HTTP server alive uri/url
# NOTE: storage embed HTTP server support uri: /status.html
http.check_alive_uri=/status.html
4> 准备client.conf配置文件
以前使用的时候只是配置了tracker.conf和storage.conf,然后就可以正常的上传文件了,但是再下载文件的时候发现就找不到了,后来才再tracker服务里面配置了client.conf才能正常的下载文件。
再下面的目录中各放一份:
/home/docker/fastdfs/tracker22122
/home/docker/fastdfs/tracker22123
/home/docker/fastdfs/tracker22124
# 你麻痹
# connect timeout in seconds
# default value is 30s
# 连接超时时间,针对socket套接字函数connect
connect_timeout=30
# network timeout in seconds
# default value is 30s
# storage server 网络超时时间,单位为秒。发送或接收数据时,如果在超时时间后还不能发送或接收数据,则本次网络通信失败。
network_timeout=60
# the base path to store log files
# 保存日志文件的位置
base_path=/fastdfs/clinet
# tracker_server can ocur more than once, and tracker_server format is
# "host:port", host can be hostname or ip address
# tracker_server 的列表 要写端口的哦 (再次提醒是主动连接tracker_server )
# 有多个tracker server时,每个tracker server写一行
tracker_server=192.168.47.150:22122
tracker_server=192.168.47.150:22123
tracker_server=192.168.47.150:22124
#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
# 日志级别
log_level=info
# if use connection pool
# default value is false
# since V4.05
# 是否使用连接池
use_connection_pool = false
# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
# 连接池最大空闲时间
connection_pool_max_idle_time = 3600
# if load FastDFS parameters from tracker server
# since V4.05
# default value is false
# 是否加载来自tracker的fdfs参数
load_fdfs_parameters_from_tracker=false
# if use storage ID instead of IP address
# same as tracker.conf
# valid only when load_fdfs_parameters_from_tracker is false
# default value is false
# since V4.05
# 是否使用存储id代替ip地址
use_storage_id = false
# specify storage ids filename, can use relative or absolute path
# same as tracker.conf
# valid only when load_fdfs_parameters_from_tracker is false
# since V4.05
storage_ids_filename = storage_ids.conf
#HTTP settings
http.tracker_server_port=80
#use "#include" directive to include HTTP other settiongs
##include http.conf
4> 准备storage.conf配置文件
该文件需要2份,但是他们配置中一个是{port=23000,group_name=group1},一个是{port=24000,group_name=group2}然后文件放在下面的目录下:
/home/docker/fastdfs/storage23000
/home/docker/fastdfs/storage24000
# 需要自行配置 /fdfs_conf/storage
# is this config file disabled
# false for enabled
# true for disabled
# 该配置文件是否不生效,false 生效;true 不生效
disabled=false
# the name of the group this storage server belongs to
# 指定 此 storage server 所在 组(卷)
group_name=group1
# bind an address of this host
# empty for bind all addresses of this host
# 是否绑定IP,后面为绑定的IP地址 (常用于服务器有多个IP但只希望一个IP提供服务)。如果不填则表示所有的
bind_addr=
# if bind an address of this host when connect to other servers
# (this storage server as a client)
# true for binding the address configed by above parameter: "bind_addr"
# false for binding any address of this host
# bind_addr通常是针对server的。当指定bind_addr时,本参数才有效。
# 本storage server作为client连接其他服务器(如tracker server、其他storage server),是否绑定bind_addr。
client_bind=true
# the storage server port
# storage server服务端口
port=23000
# connect timeout in seconds
# default value is 30s
# 连接超时时间,针对socket套接字函数connect
connect_timeout=30
# network timeout in seconds
# default value is 30s
# storage server 网络超时时间,单位为秒。发送或接收数据时,如果在超时时间后还不能发送或接收数据,则本次网络通信失败。
network_timeout=60
# heart beat interval in seconds
# 心跳间隔时间,单位为秒 (这里是指主动向tracker server 发送心跳)
heart_beat_interval=30
# disk usage report interval in seconds
# storage server向tracker server报告磁盘剩余空间的时间间隔,单位为秒。
stat_report_interval=60
# the base path to store data and log files
# base_path 目录地址,根目录必须存在 子目录会自动生成 (注 :这里不是上传的文件存放的地址,之前是的,在某个版本后更改了)
base_path=/fastdfs/storage
# max concurrent connections the server supported
# default value is 256
# more max_connections means more memory will be used
# 系统提供服务时的最大连接数。对于V1.x,因一个连接由一个线程服务,也就是工作线程数。
# 对于V2.x,最大连接数和工作线程数没有任何关系
max_connections=256
# the buff size to recv / send data
# this parameter must more than 8KB
# default value is 64KB
# since V2.00
# V2.0引入本参数。设置队列结点的buffer大小。工作队列消耗的内存大小 = buff_size * max_connections
# 设置得大一些,系统整体性能会有所提升。
# 消耗的内存请不要超过系统物理内存大小。另外,对于32位系统,请注意使用到的内存不要超过3GB
buff_size = 256KB
# accept thread count
# default value is 1
# since V4.07
accept_threads=1
# work thread count, should <= max_connections
# work thread deal network io
# default value is 4
# since V2.00
# V2.0引入的这个参数,工作线程数,通常设置为CPU数
work_threads=4
# if disk read / write separated
## false for mixed read and write
## true for separated read and write
# default value is true
# since V2.00
# V2.0引入本参数。磁盘IO读写是否分离,缺省是分离的。
disk_rw_separated = true
# disk reader thread count per store base path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
# V2.0引入本参数。针对单个存储路径的读线程数,缺省值为1。
# 读写分离时,系统中的读线程数 = disk_reader_threads * store_path_count
# 读写混合时,系统中的读写线程数 = (disk_reader_threads + disk_writer_threads) * store_path_count
disk_reader_threads = 1
# disk writer thread count per store base path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
# V2.0引入本参数。针对单个存储路径的写线程数,缺省值为1。
# 读写分离时,系统中的写线程数 = disk_writer_threads * store_path_count
# 读写混合时,系统中的读写线程数 = (disk_reader_threads + disk_writer_threads) * store_path_count
disk_writer_threads = 1
# when no entry to sync, try read binlog again after X milliseconds
# must > 0, default value is 200ms
# 同步文件时,如果从binlog中没有读到要同步的文件,休眠N毫秒后重新读取。0表示不休眠,立即再次尝试读取。
# 出于CPU消耗考虑,不建议设置为0。如何希望同步尽可能快一些,可以将本参数设置得小一些,比如设置为10ms
sync_wait_msec=50
# after sync a file, usleep milliseconds
# 0 for sync successively (never call usleep)
# 同步上一个文件后,再同步下一个文件的时间间隔,单位为毫秒,0表示不休眠,直接同步下一个文件。
sync_interval=0
# storage sync start time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
sync_start_time=00:00
# storage sync end time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# 上面二个一起解释。允许系统同步的时间段 (默认是全天) 。一般用于避免高峰同步产生一些问题而设定
sync_end_time=23:59
# write to the mark file after sync N files
# default value is 500
# 同步完N个文件后,把storage的mark文件同步到磁盘
# 注:如果mark文件内容没有变化,则不会同步
write_mark_file_freq=500
# path(disk or mount point) count, default value is 1
# 存放文件时storage server支持多个路径(例如磁盘)。这里配置存放文件的基路径数目,通常只配一个目录。
store_path_count=1
# store_path#, based 0, if store_path0 not exists, it's value is base_path
# the paths must be exist
# 逐一配置store_path个路径,索引号基于0。注意配置方法后面有0,1,2 ......,需要配置0到store_path - 1。
# 如果不配置base_path0,那边它就和base_path对应的路径一样。
store_path0=/fastdfs/store_path
#store_path1=/home/yuqing/fastdfs2
# subdir_count * subdir_count directories will be auto created under each
# store_path (disk), value can be 1 to 256, default value is 256
# FastDFS存储文件时,采用了两级目录。这里配置存放文件的目录个数 (系统的存储机制,大家看看文件存储的目录就知道了)
# 如果本参数只为N(如:256),那么storage server在初次运行时,会自动创建 N * N 个存放文件的子目录。
subdir_count_per_path=256
# tracker_server can ocur more than once, and tracker_server format is
# "host:port", host can be hostname or ip address
# tracker_server 的列表 要写端口的哦 (再次提醒是主动连接tracker_server )
# 有多个tracker server时,每个tracker server写一行
tracker_server=192.168.47.150:22122
tracker_server=192.168.47.150:22123
tracker_server=192.168.47.150:22124
#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
# 日志级别
log_level=info
#unix group name to run this program,
#not set (empty) means run by the group of current user
# 操作系统运行FastDFS的用户组 (不填 就是当前用户组,哪个启动进程就是哪个)
run_by_group=
#unix username to run this program,
#not set (empty) means run by current user
# 操作系统运行FastDFS的用户 (不填 就是当前用户,哪个启动进程就是哪个)
run_by_user=
# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" means match all ip addresses, can use range like this: 10.0.1.[1-15,20] or
# host[01-08,20-25].domain.com, for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
# 允许连接本storage server的IP地址列表 (不包括自带HTTP服务的所有连接)
# 可以配置多行,每行都会起作用
allow_hosts=*
# the mode of the files distributed to the data path
# 0: round robin(default)
# 1: random, distributted by hash code
# 文件在data目录下分散存储策略。
# 0: 轮流存放,在一个目录下存储设置的文件数后(参数file_distribute_rotate_count中设置文件数),使用下一个目录进行存储。
# 1: 随机存储,根据文件名对应的hash code来分散存储。
file_distribute_path_mode=0
# valid when file_distribute_to_path is set to 0 (round robin),
# when the written file count reaches this number, then rotate to next path
# default value is 100
# 当上面的参数file_distribute_path_mode配置为0(轮流存放方式)时,本参数有效。
# 当一个目录下的文件存放的文件数达到本参数值时,后续上传的文件存储到下一个目录中。
file_distribute_rotate_count=100
# call fsync to disk when write big file
# 0: never call fsync
# other: call fsync when written bytes >= this bytes
# default value is 0 (never call fsync)
# 当写入大文件时,每写入N个字节,调用一次系统函数fsync将内容强行同步到硬盘。0表示从不调用fsync
fsync_after_written_bytes=0
# sync log buff to disk every interval seconds
# must > 0, default value is 10 seconds
# 同步或刷新日志信息到硬盘的时间间隔,单位为秒
# 注意:storage server 的日志信息不是时时写硬盘的,而是先写内存。
sync_log_buff_interval=10
# sync binlog buff / cache to disk every interval seconds
# default value is 60 seconds
# 同步binglog(更新操作日志)到硬盘的时间间隔,单位为秒
# 本参数会影响新上传文件同步延迟时间
sync_binlog_buff_interval=10
# sync storage stat info to disk every interval seconds
# default value is 300 seconds
# 把storage的stat文件同步到磁盘的时间间隔,单位为秒。
# 注:如果stat文件内容没有变化,不会进行同步
sync_stat_file_interval=300
# thread stack size, should >= 512KB
# default value is 512KB
# 线程栈的大小。FastDFS server端采用了线程方式。
# 对于V1.x,storage server线程栈不应小于512KB;对于V2.0,线程栈大于等于128KB即可。
# 线程栈越大,一个线程占用的系统资源就越多。
# 对于V1.x,如果要启动更多的线程(max_connections),可以适当降低本参数值。
thread_stack_size=512KB
# the priority as a source server for uploading file.
# the lower this value, the higher its uploading priority.
# default value is 10
# 本storage server作为源服务器,上传文件的优先级,可以为负数。值越小,优先级越高。这里就和 tracker.conf 中store_server= 2时的配置相对应了
upload_priority=10
# the NIC alias prefix, such as eth in Linux, you can see it by ifconfig -a
# multi aliases split by comma. empty value means auto set by OS type
# default values is empty
if_alias_prefix=
# if check file duplicate, when set to true, use FastDHT to store file indexes
# 1 or yes: need check
# 0 or no: do not check
# default value is 0
# 是否检测上传文件已经存在。如果已经存在,则不存在文件内容,建立一个符号链接以节省磁盘空间。
# 这个应用要配合FastDHT 使用,所以打开前要先安装FastDHT
# 1或yes 是检测,0或no 是不检测
check_file_duplicate=0
# file signature method for check file duplicate
## hash: four 32 bits hash code
## md5: MD5 signature
# default value is hash
# since V4.01
# 文件去重时,文件内容的签名方式:
## hash: 4个hash code
## md5:MD5
file_signature_method=hash
# namespace for storing file indexes (key-value pairs)
# this item must be set when check_file_duplicate is true / on
# 当上个参数设定为1 或 yes时 (true/on也是可以的) , 在FastDHT中的命名空间
key_namespace=FastDFS
# set keep_alive to 1 to enable persistent connection with FastDHT servers
# default value is 0 (short connection)
# 与FastDHT servers 的连接方式 (是否为持久连接) ,默认是0(短连接方式)。可以考虑使用长连接,这要看FastDHT server的连接数是否够用。
keep_alive=0
# you can use "#include filename" (not include double quotes) directive to
# load FastDHT server list, when the filename is a relative path such as
# pure filename, the base path is the base path of current/this config file.
# must set FastDHT server list when check_file_duplicate is true / on
# please see INSTALL of FastDHT for detail
##include /home/yuqing/fastdht/conf/fdht_servers.conf
# 可以通过 #include filename 方式来加载 FastDHT servers 的配置,装上FastDHT就知道该如何配置啦。
# 同样要求 check_file_duplicate=1 时才有用,不然系统会忽略
# fdht_servers.conf 记载的是 FastDHT servers 列表
# if log to access log
# default value is false
# since V4.00
# 是否将文件操作记录到access log
use_access_log = false
# if rotate the access log every day
# default value is false
# since V4.00
# 是否定期轮转access log,目前仅支持一天轮转一次
rotate_access_log = false
# rotate access log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.00
# access log定期轮转的时间点,只有当rotate_access_log设置为true时有效
access_log_rotate_time=00:00
# if rotate the error log every day
# default value is false
# since V4.02
# 是否定期轮转error log,目前仅支持一天轮转一次
rotate_error_log = false
# rotate error log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.02
# error log定期轮转的时间点,只有当rotate_error_log设置为true时有效
error_log_rotate_time=00:00
# rotate access log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
# access log按文件大小轮转
# 设置为0表示不按文件大小轮转,否则当access log达到该大小,就会轮转到新文件中
rotate_access_log_size = 0
# rotate error log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
# error log按文件大小轮转
# 设置为0表示不按文件大小轮转,否则当error log达到该大小,就会轮转到新文件中
rotate_error_log_size = 0
# if skip the invalid record when sync file
# default value is false
# since V4.02
# 文件同步的时候,是否忽略无效的binlog记录
file_sync_skip_invalid_record=false
# if use connection pool
# default value is false
# since V4.05
# 是否使用连接池
use_connection_pool = false
# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
# 连接池最大空闲时间
connection_pool_max_idle_time = 3600
# use the ip address of this storage server if domain_name is empty,
# else this domain name will ocur in the url redirected by the tracker server
# storage server上web server域名,通常仅针对单独部署的web server。这样URL中就可以通过域名方式来访问storage server上的文件了,
# 这个参数为空就是IP地址的方式。
http.domain_name=
# the port of the web server on this storage server
# 如果系统较大,这个服务有可能支持不了,可以自行换一个webserver,我喜欢lighttpd,当然ng也很好了。
http.server_port=8888
5> 创建fdfs-compose.yml文件
该文件放在如下目录:
/home/docker/fastdfs
version: '3'
services:
tracker22122:
image: yxq18509376997/fastdfs
container_name: tracker22122
command: tracker
restart: always
network_mode: host
volumes:
- ./tracker22122/tracker.conf:/fdfs_conf/tracker.conf
- ./tracker22122/client.conf:/fdfs_conf/client.conf
- ./tracker22122/data:/fastdfs/tracker/data
tracker22123:
image: yxq18509376997/fastdfs
container_name: tracker22123
command: tracker
restart: always
network_mode: host
volumes:
- ./tracker22123/tracker.conf:/fdfs_conf/tracker.conf
- ./tracker22123/client.conf:/fdfs_conf/client.conf
- ./tracker22123/data:/fastdfs/tracker/data
tracker22124:
image: yxq18509376997/fastdfs
container_name: tracker22124
command: tracker
restart: always
network_mode: host
volumes:
- ./tracker22124/tracker.conf:/fdfs_conf/tracker.conf
- ./tracker22124/client.conf:/fdfs_conf/client.conf
- ./tracker22124/data:/fastdfs/tracker/data
storage23000:
image: yxq18509376997/fastdfs
container_name: storage23000
command: storage
restart: always
network_mode: host
volumes:
- ./storage23000/storage.conf:/fdfs_conf/storage.conf
- ./data/group1/M00:/fastdfs/store_path/data
- ./storage23000/data:/fastdfs/storage/data
storage24000:
image: yxq18509376997/fastdfs
container_name: storage24000
command: storage
restart: always
network_mode: host
volumes:
- ./storage24000/storage.conf:/fdfs_conf/storage.conf
- ./data/group2/M00:/fastdfs/store_path/data
- ./storage24000/data:/fastdfs/storage/data
三、启动
进入到如下目录:
cd /home/docker/fastdfs
执行如下命令:
docker-compose -f fdfs-compose.yml up -d
然后可以查看到启动了5个容器tracker22122,tracker22123,tracker22124,storage23000,storage24000
四、测试
1> 开启一个客户端
docker run -tid --name fdfs-test --net=host yxq18509376997/fastdfs sh
2> 复制拷贝client.conf配置文件
docker cp /home/docker/fastdfs/tracker22122/client.conf fdfs-test:/fdfs_conf/
3> 进入到fdfs-test容器中
docker exec -it fdfs-test /bin/bash
4> 创建一个测试txt文件
echo haha>test.txt
5> 进行文件上传测试
fdfs_upload_file /fdfs_conf/client.conf /test.txt
可能报出这样的一个错误
如果出现了这样的错误你就执行下面的命令
mkdir /fastdfs/client
之后再重新执行上传命令,然后就可以看到
6> 下载文件
fdfs_download_file /fdfs_conf/client.conf group1/M00/00/00/wKgvll2FvIOAAejZAAAABRvwYuU534.txt
然后你就可以看到
7> 删除文件
dfs_delete_file /fdfs_conf/client.conf group1/M00/00/00/wKgvll2FvIOAAejZAAAABRvwYuU534.txt
然后查看文件原来所在的位置
文件已经被删除了。
注:java操作fastdfs进行上传,下载,删除,设置文件元信息,分组上传缩略图等可以查看网络上关于fastdfs-client.jar包的相关操作,或联系本人[email protected]