单机FastDFS基本用法

什么是FastDFS

FastDFS是一个开源的轻量级分布式文件系统,它对文件进行管理,功能包括:文件存储、文件同步、文件访问(文件上传、文件下载)等,解决了大容量存储和负载均衡的问题。特别适合以文件(建议范围:4KB < file_size <500MB)为载体的在线服务,如相册网站、视频网站等等。

FastDFS架构

在这里插入图片描述

FastDFS角色

  • tracker server:调度器,负责维持集群的信息,管理着所有的group和storage server。每个storage在启动之后都会连接tracker,并告知tracker自己的心跳信息。 tracker不存储数据,它管理的元数据信息全部放置在内存中,并且元数据信息是根据storage汇报的信息生成,本身不做持久化数据。
  • storage server:存储服务器,以group为单位,一个group中有多个存储服务器,在同一个group中,存储服务器中的数据互相同步。 每一个storage的存储对应一个目录,比如,有10个硬盘,可以设置每个硬盘的对应的目录为/data/dev1…/data/dev10。 storage在目录中存储数据的时候,会创建2级子目录,每级256个,新写的文件会以hash的方式被路由到其中某个子目录下。
  • client:客户端,用来上传,下载,追加,删除等数据。

上传文件

  • 由client发起上传连接请求
  • 由选定的tracker查找可以用的storage server
  • 找到可用的storage server后,将其返回给客户端
  • 上传文件
  • 生成文件的fid,将client的内容写入选定的位置
  • 返回fid
  • 同步数据到同组内的其它节点

tracker挑选组的方式

  • round robin
  • 指定组
  • 基于可用空间进行挑选,剩余可用空间大的优先

group内挑选storage server

  • round robin
  • 以ip次序,找第一个
  • 以优先级排序来找

如何选择磁盘

  • round robin
  • 剩余可用空间大优先

FID:由源头的stroage server ip,创建时间戳,大小,文件的检验码和一个随机数进行hash计算后生成,最后基于base64编码,转换为可打印字符
文件名:当文件存储到某个子目录后,即认为该文件存储成功,接下来会为该文件生成一个文件名,文件名由group、存储目录、两级子目录、fileid、文件后缀名(由客户端指定,主要用于区分文件类型)拼接而成。

文件下载

客户端上传完文件后,会接收到服务器返回的fid,而后在次用到时,client可以根据此文件发出请求

  • client向tarcker发起请求
  • tracker根据文件名定位到group,并返回group内的某一个storage server 的信息给client
  • client根据返回的信息发请求
  • storage server查找文件,并返回给client。

FastDFS使用

文档介绍

1)安装编译环境

扫描二维码关注公众号,回复: 12278412 查看本文章
	yum install git gcc gcc-c++ make automake autoconf libtool pcre pcre-devel zlib zlib-devel openssl-devel wget vim -y

2)安装包路径

[root@ydong src]# pwd
/usr/local/src

3)编译安装libfastcommonfastdfs

[root@ydong libfastcommon-master]# pwd
/usr/local/src/libfastcommon-master
./make.sh && ./make.sh install 

[root@ydong src]# cd fastdfs-master/
[root@ydong fastdfs-master]# pwd
/usr/local/src/fastdfs-master

./make.sh && ./make.sh install

4)配置文件目录

[root@ydong fdfs]# pwd
/etc/fdfs

[root@ydong fdfs]# ls
client.conf.sample  storage.conf.sample  storage_ids.conf.sample  tracker.conf.sample

client.conf:客户端测试使用
storage.conf:存储节点配置
tracker.conf:tracker节点配置

5)trakcer.conf配置文件介绍

# is this config file disabled
# false for enabled
# true for disabled
disabled = false   是否启用配置文件,false表示启用

# bind an address of this host
# empty for bind all addresses of this host
bind_addr =       绑定地址,不写表示绑定所有的地址
 
# the tracker server port
port = 22122    绑定端口

# connect timeout in seconds
# default value is 30
# Note: in the intranet network (LAN), 2 seconds is enough.
connect_timeout = 5       连接超时时长,在内部网,2s足够

# network timeout in seconds for send and recv
# default value is 30
network_timeout = 60     网络连接超时时长

# the base path to store data and log files
base_path = /home/yuqing/fastdfs   数据存储目录和日志目录

# max concurrent connections this server support
# you should set this parameter larger, eg. 10240   
# default value is 256
max_connections = 1024    最大并发连接数,建议设置大一些,例如10240

# accept thread count
# default value is 1 which is recommended
# since V4.07
accept_threads = 1    接收的线程数量,推荐1

# work thread count
# work threads to deal network io
# default value is 4
# since V2.00
work_threads = 4    工作的线程池数量,应该小于等于cpu核心数

# the min network buff size
# default value 8KB
min_buff_size = 8KB   设置网络缓冲

# the max network buff size
# default value 128KB
max_buff_size = 128KB

# the method for selecting group to upload files
# 0: round robin   轮询
# 1: specify group   指定组名
# 2: load balance, select the max free space group to upload file   选择最大的可用剩余空间组
store_lookup = 2   设定上传文件的方法

# which group to upload file
# when store_lookup set to 1, must set store_group to the group name
store_group = group2   如果上面store_lookup = 1 ,此处需要写上store_group的名字

# which storage server to upload file
# 0: round robin (default)   轮询
# 1: the first server order by ip address    指定的ip地址
# 2: the first server order by priority (the minimal)      按照优先级排序  
# Note: if use_trunk_file set to true, must set store_server to 1 or 2
store_server = 0   设定storage server上传文件的方法

# which path (means disk or mount point) of the storage server to upload file
# 0: round robin  轮询
# 2: load balance, select the max free space path to upload file  选择最大可用剩余空间
store_path = 0   选择storage server上的哪个存储路径

# which storage server to download file
# 0: round robin (default)   轮询
# 1: the source storage server which the current file uploaded to    文件第一次上传到哪个,就从哪个下载
download_server = 0    使用哪个storage server下载文件

# reserved storage space for system or other applications.
# if the free(available) space of any stoarge server in 
# a group <= reserved_storage_space, no file can be uploaded to this group.
# bytes unit can be one of follows:
### G or g for gigabyte(GB)
### M or m for megabyte(MB)
### K or k for kilobyte(KB)
### no unit for byte(B)
### XX.XX% as ratio such as: reserved_storage_space = 10%
reserved_storage_space = 20%    设定组内可剩余的存储空间,如果低于等于这个数,就不能存储到这个组上去

#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level = info   设定存储日志的级别

#unix group name to run this program, 
#not set (empty) means run by the group of current user
run_by_group=   设定运行的组名
 
#unix username to run this program,
#not set (empty) means run by current user
run_by_user =     设定运行的用户  空表示当前用户

# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" (only one asterisk) means match all ip addresses
# we can use CIDR ips like 192.168.5.64/26
# and also use range like these: 10.0.1.[0-254] and host[01-08,20-25].domain.com
# for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
# allow_hosts=192.168.5.64/26
allow_hosts = *   允许哪些主机连接

# sync log buff to disk every interval seconds
# default value is 10 seconds
sync_log_buff_interval = 1  同步日志每隔几秒从缓冲写到磁盘中

# check storage server alive interval seconds
check_active_interval = 120    每隔多长时间检测一次storage server的心跳

# thread stack size, should >= 64KB
# default value is 256KB
thread_stack_size = 256KB

# auto adjust when the ip address of the storage server changed
# default value is true
storage_ip_changed_auto_adjust = true   当storage server的ip地址变动时,自动调整

# storage sync file max delay seconds
# default value is 86400 seconds (one day)
# since V2.00
storage_sync_file_max_delay = 86400   存储服务器同步文件最大延迟数

# the max time of storage sync a file
# default value is 300 seconds
# since V2.00
storage_sync_file_max_time = 300   存储服务器同步一个文件需要消耗的时间

# if use a trunk file to store several small files
# default value is false
# since V3.00
use_trunk_file = false  是否将小文件合并成一个大文件

# the min slot size, should <= 4KB
# default value is 256 bytes
# since V3.00
slot_min_size = 256   trunk file分配的最小字节数

# the max slot size, should > slot_min_size
# store the upload file to trunk file when it's size <=  this value
# default value is 16MB
# since V3.00
slot_max_size = 1MB   trunk file分配的最大字节数

# the alignment size to allocate the trunk space
# default value is 0 (never align)
# since V6.05
# NOTE: the larger the alignment size, the less likely of disk
#       fragmentation, but the more space is wasted.
trunk_alloc_alignment_size = 256

# if merge contiguous free spaces of trunk file
# default value is false
# since V6.05
trunk_free_space_merge = true

# if delete / reclaim the unused trunk files
# default value is false
# since V6.05
delete_unused_trunk_files = false

# the trunk file size, should >= 4MB
# default value is 64MB
# since V3.00
trunk_file_size = 64MB   设定trunk file的大小

# if create trunk file advancely
# default value is false
# since V3.06
trunk_create_file_advance = false  是否提前创建trunk file

# the time base to create trunk file
# the time format: HH:MM
# default value is 02:00
# since V3.06
trunk_create_file_time_base = 02:00   创建trunk file的时间

# the interval of create trunk file, unit: second
# default value is 38400 (one day)
# since V3.06
trunk_create_file_interval = 86400   创建trunk file的时间间隔,单位为秒。假设每天仅仅提前创建一次,则设置为86400

# the threshold to create trunk file
# when the free trunk file size less than the threshold,
# will create he trunk files
# default value is 0
# since V3.06
trunk_create_file_space_threshold = 20G   创建trunk file的阈值

# if check trunk space occupying when loading trunk free spaces
# the occupied spaces will be ignored
# default value is false
# since V3.09
# NOTICE: set this parameter to true will slow the loading of trunk spaces 
# when startup. you should set this parameter to true when neccessary.
trunk_init_check_occupying = false   trunk初始化时,是否检查可用空间是否被占用

# if ignore storage_trunk.dat, reload from trunk binlog
# default value is false
# since V3.10
# set to true once for version upgrade when your version less than V3.10
trunk_init_reload_from_binlog = false   是否无条件从trunk binlog中载入trunk可用空间信息

# the min interval for compressing the trunk binlog file
# unit: second, 0 means never compress
# FastDFS compress the trunk binlog when trunk init and trunk destroy
# recommand to set this parameter to 86400 (one day)
# default value is 0
# since V5.01
trunk_compress_binlog_min_interval = 86400    压缩trunk binlog文件的最小间隔时长

# the interval for compressing the trunk binlog file
# unit: second, 0 means never compress
# recommand to set this parameter to 86400 (one day)
# default value is 0
# since V6.05
trunk_compress_binlog_interval = 86400   压缩trunk binlog文件的间隔时长


# compress the trunk binlog time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 03:00
# since V6.05
trunk_compress_binlog_time_base = 03:00   在什么时候压缩trunk binlog

# max backups for the trunk binlog file
# default value is 0 (never backup)
# since V6.05
trunk_binlog_max_backups = 7  备份trunk binlog的最大份数

# if use storage server ID instead of IP address
# if you want to use dual IPs for storage server, you MUST set
# this parameter to true, and configure the dual IPs in the file
# configured by following item "storage_ids_filename", such as storage_ids.conf
# default value is false
# since V4.00
use_storage_id = false   是否使用storage server ID作为标识

# specify storage ids filename, can use relative or absolute path
# this parameter is valid only when use_storage_id set to true
# since V4.00
storage_ids_filename = storage_ids.conf   use_storage_id 设置为true,才须要设置本參数

# id type of the storage server in the filename, values are:
## ip: the ip address of the storage server
## id: the server id of the storage server
# this paramter is valid only when use_storage_id set to true
# default value is ip
# since V4.03
id_type_in_filename = id   文件名称中的id类型,有ip和id两种,仅仅有当use_storage_id设置为true时该參数才有效

# if store slave file use symbol link
# default value is false
# since V4.01
store_slave_file_use_link = false    存储文件是否使用符号链接方式

# if rotate the error log every day
# default value is false
# since V4.02
rotate_error_log = false  是否每天轮换错误日志

# rotate error log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.02
error_log_rotate_time = 00:00   轮换错误日志的时间

# if compress the old error log by gzip
# default value is false
# since V6.04
compress_old_error_log = false   是否压缩错误日志

# compress the error log days before
# default value is 1
# since V6.04
compress_error_log_days_before = 7  压缩几天前的错误日志

# rotate error log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_error_log_size = 0  当日志文件超过此处设定的大小,就翻滚此日志

# keep days of the log files
# 0 means do not delete old log files
# default value is 0
log_file_keep_days = 0   保存日志的天数,0表示不删除旧日志
 
# if use connection pool
# default value is false
# since V4.05
use_connection_pool = true   是否使用连接池

# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
connection_pool_max_idle_time = 3600     假设一个连接的空暇时间超过这个值将会被自己主动关闭

# HTTP port on this tracker server
http.server_port = 8080   http server端口

# check storage HTTP server alive interval seconds
# <= 0 for never check
# default value is 30
http.check_alive_interval = 30   检测健康状态间隔

# check storage HTTP server alive type, values are:
#   tcp : connect to the storge server with HTTP port only, 
#        do not request and get response
#   http: storage check alive url must return http status 200
# default value is tcp
http.check_alive_type = tcp   使用什么方式检测

# check storage HTTP server alive uri/url
# NOTE: storage embed HTTP server support uri: /status.html
http.check_alive_uri = /status.html    检测的uri

6)修改tracker配置,启动tracker

只修改了一处
base_path = /data/fastdfs

[root@ydong fdfs]# mkdir /data/fastdfs -pv
mkdir: 已创建目录 "/data"
mkdir: 已创建目录 "/data/fastdfs"

7)启动tracker节点

[root@ydong fdfs]# /etc/init.d/fdfs_trackerd  start
Reloading systemd:                                         [  确定  ]
Starting fdfs_trackerd (via systemctl):                    [  确定  ]

[root@ydong fdfs]# ss -tnlp | grep 22122
LISTEN     0      128          *:22122                    *:*                   users:(("fdfs_trackerd",pid=1326,fd=5))

8)storage.conf配置文件

# is this config file disabled
# false for enabled
# true for disabled
disabled = false   是否启用配置

# the name of the group this storage server belongs to
#
# comment or remove this item for fetching from tracker server,
# in this case, use_storage_id must set to true in tracker.conf,
# and storage_ids.conf must be configured correctly.
group_name = group1  属于哪个组

# bind an address of this host
# empty for bind all addresses of this host
bind_addr =   绑定地址

# if bind an address of this host when connect to other servers 
# (this storage server as a client)
# true for binding the address configured by the above parameter: "bind_addr"
# false for binding any address of this host
client_bind = true       连接其他服务器时绑定的地址,true表示以bind_addr设定的,false表示任意地址

# the storage server port
port = 23000  端口

# connect timeout in seconds
# default value is 30
# Note: in the intranet network (LAN), 2 seconds is enough.
connect_timeout = 5   连接超时时长

# network timeout in seconds for send and recv
# default value is 30
network_timeout = 60   网络连接超时时长

# the heart beat interval in seconds
# the storage server send heartbeat to tracker server periodically
# default value is 30
heart_beat_interval = 30   storage server向tracker发送心跳的间隔时长

# disk usage report interval in seconds
# the storage server send disk usage report to tracker server periodically
# default value is 300
stat_report_interval = 60  每隔多长时间向tracker发送磁盘使用情况

# the base path to store data and log files
# NOTE: the binlog files maybe are large, make sure
#       the base path has enough disk space,
#       eg. the disk free space should > 50GB
base_path = /home/yuqing/fastdfs    storage server的数据存储路径和日志路径

# max concurrent connections the server supported,
# you should set this parameter larger, eg. 10240
# default value is 256
max_connections = 1024  最大并发连接数

# the buff size to recv / send data from/to network
# this parameter must more than 8KB
# 256KB or 512KB is recommended
# default value is 64KB
# since V2.00
buff_size = 256KB

# accept thread count
# default value is 1 which is recommended
# since V4.07
accept_threads = 1

# work thread count
# work threads to deal network io
# default value is 4
# since V2.00
work_threads = 4

# if disk read / write separated
##  false for mixed read and write
##  true for separated read and write
# default value is true
# since V2.00
disk_rw_separated = true  是否读写分离

# disk reader thread count per store path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
disk_reader_threads = 1  每个存储基路径的磁盘读取器线程计数

# disk writer thread count per store path
# for mixed read / write, this parameter can be 0
# default value is 1
# since V2.00
disk_writer_threads = 1   每个存储基路径的磁盘写入器线程计数

# when no entry to sync, try read binlog again after X milliseconds
# must > 0, default value is 200ms
sync_wait_msec = 50  当没有同步条目时,在多久之后尝试重新读取binlog

# after sync a file, usleep milliseconds
# 0 for sync successively (never call usleep)
sync_interval = 0  在同步完文件后,休息多长时间

# storage sync start time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
sync_start_time = 00:00  同步的时间

# storage sync end time of a day, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
sync_end_time = 23:59  同步结束的时间

# write to the mark file after sync N files
# default value is 500
write_mark_file_freq = 500  同步完N个文件后,把storage的mark文件同步到磁盘

# disk recovery thread count
# default value is 1
# since V6.04
disk_recovery_threads = 3   磁盘恢复线程数量

# store path (disk or mount point) count, default value is 1
store_path_count = 1    存放文件时storage server支持多个路径(比如磁盘)。这里配置存放文件的基路径数目。通常仅仅配一个文件夹。

# store_path#, based on 0, to configure the store paths to store files
# if store_path0 not exists, it's value is base_path (NOT recommended)
# the paths must be exist.
#
# IMPORTANT NOTE:
#       the store paths' order is very important, don't mess up!!!
#       the base_path should be independent (different) of the store paths

store_path0 = /home/yuqing/fastdfs    store_path0的路径,如果上面store_path_count=2,则需要在此处加上`store_path1="PATH"`,以此类推
#store_path1 = /home/yuqing/fastdfs2

# subdir_count  * subdir_count directories will be auto created under each 
# store_path (disk), value can be 1 to 256, default value is 256
subdir_count_per_path = 256  自动创建子目录

# tracker_server can ocur more than once for multi tracker servers.
# the value format of tracker_server is "HOST:PORT",
#   the HOST can be hostname or ip address,
#   and the HOST can be dual IPs or hostnames seperated by comma,
#   the dual IPS must be an inner (intranet) IP and an outer (extranet) IP,
#   or two different types of inner (intranet) IPs.
#   for example: 192.168.2.100,122.244.141.46:22122
#   another eg.: 192.168.1.10,172.17.4.21:22122

tracker_server = 192.168.209.121:22122   tracker服务器的地址和端口,可以出现多个
tracker_server = 192.168.209.122:22122

#standard log level as syslog, case insensitive, value list:
### emerg for emergency
### alert
### crit for critical
### error
### warn for warning
### notice
### info
### debug
log_level = info  日志级别

#unix group name to run this program, 
#not set (empty) means run by the group of current user
run_by_group =

#unix username to run this program,
#not set (empty) means run by current user
run_by_user =

# allow_hosts can ocur more than once, host can be hostname or ip address,
# "*" (only one asterisk) means match all ip addresses
# we can use CIDR ips like 192.168.5.64/26
# and also use range like these: 10.0.1.[0-254] and host[01-08,20-25].domain.com
# for example:
# allow_hosts=10.0.1.[1-15,20]
# allow_hosts=host[01-08,20-25].domain.com
# allow_hosts=192.168.5.64/26
allow_hosts = *

# the mode of the files distributed to the data path
# 0: round robin(default)  轮询
# 1: random, distributted by hash code    随机,由hash分配
file_distribute_path_mode = 0   文件在磁盘路径下的分发目录

# valid when file_distribute_to_path is set to 0 (round robin).
# when the written file count reaches this number, then rotate to next path.
# rotate to the first path (00/00) after the last path (such as FF/FF).
# default value is 100
file_distribute_rotate_count = 100   当上面的參数file_distribute_path_mode配置为0(轮流存放方式)时,本參数有效。 当一个文件夹下的文件存放的文件数达到本參数值时,兴许上传的文件存储到下一个文件夹中。

# call fsync to disk when write big file
# 0: never call fsync
# other: call fsync when written bytes >= this bytes
# default value is 0 (never call fsync)
fsync_after_written_bytes = 0  当写入一个大文件时,调用fsync方式写入磁盘

# sync log buff to disk every interval seconds
# must > 0, default value is 10 seconds
sync_log_buff_interval = 1  间隔多长时间从缓冲写入到磁盘

# sync binlog buff / cache to disk every interval seconds
# default value is 60 seconds
sync_binlog_buff_interval = 1   间隔多长时间将binlog buff

# sync storage stat info to disk every interval seconds
# default value is 300 seconds
sync_stat_file_interval = 300  每隔多长时间将数据存储信息同步到磁盘上

# thread stack size, should >= 512KB
# default value is 512KB
thread_stack_size = 512KB

# the priority as a source server for uploading file.
# the lower this value, the higher its uploading priority.
# default value is 10
upload_priority = 10   storage的优先级

# the NIC alias prefix, such as eth in Linux, you can see it by ifconfig -a
# multi aliases split by comma. empty value means auto set by OS type
# default values is empty
if_alias_prefix =

# if check file duplicate, when set to true, use FastDHT to store file indexes
# 1 or yes: need check
# 0 or no: do not check
# default value is 0
check_file_duplicate = 0   检查文件是否重复,1表示检查,则由fastDHT索引来存储

# file signature method for check file duplicate
## hash: four 32 bits hash code
## md5: MD5 signature
# default value is hash
# since V4.01
file_signature_method = hash   检查文件签名

# namespace for storing file indexes (key-value pairs)
# this item must be set when check_file_duplicate is true / on
key_namespace = FastDFS   存储文件的索引

# set keep_alive to 1 to enable persistent connection with FastDHT servers
# default value is 0 (short connection)
keep_alive = 0   设定与FastDHT的持久连接,0表示短连接

# you can use "#include filename" (not include double quotes) directive to 
# load FastDHT server list, when the filename is a relative path such as 
# pure filename, the base path is the base path of current/this config file.
# must set FastDHT server list when check_file_duplicate is true / on
# please see INSTALL of FastDHT for detail
##include /home/yuqing/fastdht/conf/fdht_servers.conf

# if log to access log
# default value is false
# since V4.00
use_access_log = false  是否启用access log

# if rotate the access log every day
# default value is false
# since V4.00
rotate_access_log = false  是否滚动access_log

# rotate access log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.00
access_log_rotate_time = 00:00   滚动日志的时间

# if compress the old access log by gzip
# default value is false
# since V6.04
compress_old_access_log = false  是否压缩旧日志

# compress the access log days before
# default value is 1
# since V6.04
compress_access_log_days_before = 7   压缩几天前的日志 

# if rotate the error log every day
# default value is false
# since V4.02
rotate_error_log = false   是否滚动error_log

# rotate error log time base, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 00:00
# since V4.02
error_log_rotate_time = 00:00   滚动error_log的时间

# if compress the old error log by gzip
# default value is false
# since V6.04
compress_old_error_log = false

# compress the error log days before
# default value is 1
# since V6.04
compress_error_log_days_before = 7

# rotate access log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_access_log_size = 0

# rotate error log when the log file exceeds this size
# 0 means never rotates log file by log file size
# default value is 0
# since V4.02
rotate_error_log_size = 0

# keep days of the log files
# 0 means do not delete old log files
# default value is 0
log_file_keep_days = 0  保存日志文件的天数

# if skip the invalid record when sync file
# default value is false
# since V4.02
file_sync_skip_invalid_record = false 是否 当同步文件时跳过无效的记录

# if use connection pool
# default value is false
# since V4.05
use_connection_pool = true

# connections whose the idle time exceeds this time will be closed
# unit: second
# default value is 3600
# since V4.05
connection_pool_max_idle_time = 3600

# if compress the binlog files by gzip
# default value is false
# since V6.01
compress_binlog = true

# try to compress binlog time, time format: Hour:Minute
# Hour from 0 to 23, Minute from 0 to 59
# default value is 01:30
# since V6.01
compress_binlog_time = 01:30

# if check the mark of store path to prevent confusion
# recommend to set this parameter to true
# if two storage servers (instances) MUST use a same store path for
# some specific purposes, you should set this parameter to false
# default value is true
# since V6.03
check_store_path_mark = true

# use the ip address of this storage server if domain_name is empty,
# else this domain name will ocur in the url redirected by the tracker server
http.domain_name =       storage server上web server域名。通常仅针对单独部署的web server。这样URL中就能够通过域名方式来訪问storage server上的文件了   这个參数为空就是IP地址的方式。

# the port of the web server on this storage server
http.server_port = 8888

9)修改storage配置文件,启动storage

base_path = /ydong/fastdfs

store_path0 = /data/fastdfs

tracker_server = 192.168.199.155:22122

[root@ydong fdfs]# mkdir /ydong/fastdfs -pv
mkdir: 已创建目录 "/ydong"
mkdir: 已创建目录 "/ydong/fastdfs"

[root@ydong fdfs]# /etc/init.d/
fdfs_storaged  fdfs_trackerd  netconsole     network        
[root@ydong fdfs]# /etc/init.d/fdfs_storaged start
Starting fdfs_storaged (via systemctl):                    [  确定  ]

[root@ydong fdfs]# ss -tnlp | grep 23000
LISTEN     0      128          *:23000                    *:*                   users:(("fdfs_storaged",pid=2205,fd=5))

10)client.conf修改配置

base_path = /ydong/client/fastdfs
tracker_server = 192.168.199.155:22122

[root@ydong fdfs]# mkdir /ydong/client/fastdfs -pv
mkdir: 已创建目录 "/ydong/client"
mkdir: 已创建目录 "/ydong/client/fastdfs"

11)上传文件测试
fdfs_upload_file <config_file> <local_filename> [storage_ip:port] [store_path_index]

[root@ydong fdfs]# fdfs_upload_file  ./client.conf /etc/fstab  
group1/M00/00/00/wKjHm1_aKYWAcK-zAAACHW1f-yE5058476

文件名组成:
group1:组名
M00:磁盘路径
00/00:1级目录/2级目录
wKj...476:hash码


12)查看文件信息
fdfs_file_info <config_file> <file_id>

[root@ydong fdfs]# fdfs_file_info  ./client.conf  group1/M00/00/00/wKjHm1_aKYWAcK-zAAACHW1f-yE5058476
GET FROM SERVER: false

file type: normal
source storage id: 0
source ip address: 192.168.199.155
file create timestamp: 2020-12-16 23:36:37
file size: 541
file crc32: 1835006753 (0x6d5ffb21)

13)下载文件
fdfs_download_file <config_file> <file_id> [local_filename] [<download_offset> <download_bytes>]

fdfs_download_file ./client.conf group1/M00/00/00/wKjHm1_aKYWAcK-zAAACHW1f-yE5058476  /root/fdfs_download 

[root@ydong fdfs]# cat /root/fdfs_download 
#
# /etc/fstab
...
/dev/mapper/centos-swap swap                    swap    defaults        0 0

上述仅仅是在fastdfs单机上测试。

猜你喜欢

转载自blog.csdn.net/qq_44564366/article/details/111244483
今日推荐