diff mbox series

[v11,2/2] zonefs: Add documentation

Message ID 20200205120837.67798-3-damien.lemoal@wdc.com (mailing list archive)
State Superseded, archived
Headers show
Series New zonefs file system | expand

Commit Message

Damien Le Moal Feb. 5, 2020, 12:08 p.m. UTC
Add the new file Documentation/filesystems/zonefs.txt to document
zonefs principles and user-space tool usage.

Signed-off-by: Damien Le Moal <damien.lemoal@wdc.com>
---
 Documentation/filesystems/zonefs.txt | 310 +++++++++++++++++++++++++++
 MAINTAINERS                          |   1 +
 2 files changed, 311 insertions(+)
 create mode 100644 Documentation/filesystems/zonefs.txt

Comments

Dave Chinner Feb. 5, 2020, 10:29 p.m. UTC | #1
On Wed, Feb 05, 2020 at 09:08:37PM +0900, Damien Le Moal wrote:
> Add the new file Documentation/filesystems/zonefs.txt to document
> zonefs principles and user-space tool usage.
> 
> Signed-off-by: Damien Le Moal <damien.lemoal@wdc.com>
.....

Just looking at the error handling text...

> +Several optional features of zonefs can be enabled at format time.
> +* Conventional zone aggregation: ranges of contiguous conventional zones can be
> +  aggregated into a single larger file instead of the default one file per zone.
> +* File ownership: The owner UID and GID of zone files is by default 0 (root)
> +  but can be changed to any valid UID/GID.
> +* File access permissions: the default 640 access permissions can be changed.
> +
> +zonefs mount options
> +--------------------

This section is really all about error handling, not so much mount
options in general...

> +
> +zonefs defines several mount options allowing the user to control the file
> +system behavior when write I/O errors occur and when inconsistencies between a
> +file size and its zone write pointer position are discovered. The handling of
> +read I/O errors is not changed by these options as long as no inode size
> +corruption is detected.
> +
> +These options are as follows.
> +* errors=remount-ro (default)
> +  All write IO errors and errors due to a zone of the device going "bad"
> +  (condition changed to offline or read-only), the file system is remounted
> +  read-only after fixing the size and access permissions of the inode that
> +  suffered the IO error.

What does "fixing the size and access permissions of the inode"
mean?

> +* errors=zone-ro
> +  Any write IO error to a file zone result in the zone being considered as in a
> +  read-only condition, preventing any further modification to the file. This
> +  option does not affect the handling of errors due to offline zones. For these
> +  zones, all accesses (read and write) are disabled.

If the zone is marked RO, then shouldn't reads still work?. Oh, hold
on, you're now talking about errors that take the zone oflfine at
the device level?

Perhaps a table describing what IO can be done to a zone vs the
device once an error occurs would be a clearer way of describing the
behaviour.


It seems to me that a table might be a better way of decribing all
the different conditions

				Post error access permissions
				   zone		    device
mountopt	zone state	read	write	read	write
--------	----------	----	-----	----	-----
remount-ro	good		yes	no	yes	no
		RO		yes	no	yes	no
		Offline		no	no	yes	no

zone-ro		good		yes	no	yes	yes
		RO		yes	no	yes	yes
		Offline		no	no	yes	yes

zone-offline	good		no	no	yes	yes
		RO		no	no	yes	yes
		Offline		no	no	yes	yes

repair		good		yes	yes	yes	yes
		RO		yes	no	yes	yes
		Offline		no	no	yes	yes

And then you can document that an offline zone will always appear to
have a size of 0, be immutable, etc, while a read-only zone will
have a size that reflects the amount of valid data in the zone that
can be read.

IOWs, you don't need to mix the definitions of zone state appearence
and behaviour with descriptions of what actions the different mount
options take when a write error occurs.

> +* errors=zone-offline
> +  Any write IO error to a file zone result in the zone being considered as in
> +  an offline condition. This implies that the file size is changed to 0 and all
> +  read/write accesses to the file disabled, preventing all accesses by the user.
> +* errors=repair
> +  Any inconsistency between an inode size and its zone amount of written data
> +  due to IO errors or external corruption are fixed without any change to file
> +  access rights. This option does not affect the processing of zones that were
> +  signaled as read-only or offline by the device. For read-only zones, the file
> +  read accesses are disabled and for offline zones, all access permissions are
> +  removed.
> +
> +For sequential zone files, inconsistencies between an inode size and the amount
> +of data writen in its zone, that is, the position of the file zone write
> +pointer, can result from different events:
> +* When the device write cache is enabled, a differed write error can occur

"a different write error"?

> +  resulting in the amount of data written in the zone being less than the inode
> +  size.

Though I suspect that what you really mean to say is that errors can
occur in ranges of previously completed writes can occur when the
cache is flushed, hence less data being physically written than the
OS has previously be told was written by the hardware. i.e. visible
inode size goes backwards.

> +* Partial failures of large write I/O operations (e.g. one BIO of a multi-bio
> +  large direct write fails) can result in the amount of data written in the
> +  zone being larger than the inode size.
> +* External action on the disk such as write, zone reset or zone finish
> +  operations will change a file zone write pointer position resulting in a
> +  reported amount of written data being different from the file inode size.

*nod*

> +Finally, defective drives may change the condition of any zone to offline (zone
> +dead) or read-only. Such changes, when discovered with the IO errors they can
> +cause, are handled automatically regardless of the options specified at mount
> +time. For offline zones, the action taken is similar to the action defined by
> +the errors=zone-offline mount option. For read-only zones, the action used is
> +as defined by the errors=zone-ro mount option.

Hmmmm. I think that's over-complicating things and takes control of
error handling away from the user. That is, regardless of the reason
for the error, if we get a write error and the user specified
errors=remount-ro, the entire device should go read-only because
that's what the user has told the filesystem to do on write error.

This seems pretty user-unfriendly - giving them a way to control
error handling behaviour and then ignoring it for specific errors
despite the fact they mean exactly the same thing to the user: the
write failed because a zone has gone bad since the last time that
zone was accessed by the application....

Cheers,

Dave.
Damien Le Moal Feb. 6, 2020, 1:42 a.m. UTC | #2
On 2020/02/06 7:29, Dave Chinner wrote:
> On Wed, Feb 05, 2020 at 09:08:37PM +0900, Damien Le Moal wrote:
>> Add the new file Documentation/filesystems/zonefs.txt to document
>> zonefs principles and user-space tool usage.
>>
>> Signed-off-by: Damien Le Moal <damien.lemoal@wdc.com>
> .....
> 
> Just looking at the error handling text...
> 
>> +Several optional features of zonefs can be enabled at format time.
>> +* Conventional zone aggregation: ranges of contiguous conventional zones can be
>> +  aggregated into a single larger file instead of the default one file per zone.
>> +* File ownership: The owner UID and GID of zone files is by default 0 (root)
>> +  but can be changed to any valid UID/GID.
>> +* File access permissions: the default 640 access permissions can be changed.
>> +
>> +zonefs mount options
>> +--------------------
> 
> This section is really all about error handling, not so much mount
> options in general...

Indeed. Section title fix is needed.

> 
>> +
>> +zonefs defines several mount options allowing the user to control the file
>> +system behavior when write I/O errors occur and when inconsistencies between a
>> +file size and its zone write pointer position are discovered. The handling of
>> +read I/O errors is not changed by these options as long as no inode size
>> +corruption is detected.
>> +
>> +These options are as follows.
>> +* errors=remount-ro (default)
>> +  All write IO errors and errors due to a zone of the device going "bad"
>> +  (condition changed to offline or read-only), the file system is remounted
>> +  read-only after fixing the size and access permissions of the inode that
>> +  suffered the IO error.
> 
> What does "fixing the size and access permissions of the inode"
> mean?
> 
>> +* errors=zone-ro
>> +  Any write IO error to a file zone result in the zone being considered as in a
>> +  read-only condition, preventing any further modification to the file. This
>> +  option does not affect the handling of errors due to offline zones. For these
>> +  zones, all accesses (read and write) are disabled.
> 
> If the zone is marked RO, then shouldn't reads still work?. Oh, hold
> on, you're now talking about errors that take the zone oflfine at
> the device level?
> 
> Perhaps a table describing what IO can be done to a zone vs the
> device once an error occurs would be a clearer way of describing the
> behaviour.
> 
> 
> It seems to me that a table might be a better way of decribing all
> the different conditions
> 
> 				Post error access permissions
> 				   zone		    device
> mountopt	zone state	read	write	read	write
> --------	----------	----	-----	----	-----
> remount-ro	good		yes	no	yes	no
> 		RO		yes	no	yes	no
> 		Offline		no	no	yes	no
> 
> zone-ro		good		yes	no	yes	yes
> 		RO		yes	no	yes	yes
> 		Offline		no	no	yes	yes
> 
> zone-offline	good		no	no	yes	yes
> 		RO		no	no	yes	yes
> 		Offline		no	no	yes	yes
> 
> repair		good		yes	yes	yes	yes
> 		RO		yes	no	yes	yes
> 		Offline		no	no	yes	yes
> 
> And then you can document that an offline zone will always appear to
> have a size of 0, be immutable, etc, while a read-only zone will
> have a size that reflects the amount of valid data in the zone that
> can be read.
> 
> IOWs, you don't need to mix the definitions of zone state appearence
> and behaviour with descriptions of what actions the different mount
> options take when a write error occurs.

Excellent idea ! That will clarify things a lot.

>> +* errors=zone-offline
>> +  Any write IO error to a file zone result in the zone being considered as in
>> +  an offline condition. This implies that the file size is changed to 0 and all
>> +  read/write accesses to the file disabled, preventing all accesses by the user.
>> +* errors=repair
>> +  Any inconsistency between an inode size and its zone amount of written data
>> +  due to IO errors or external corruption are fixed without any change to file
>> +  access rights. This option does not affect the processing of zones that were
>> +  signaled as read-only or offline by the device. For read-only zones, the file
>> +  read accesses are disabled and for offline zones, all access permissions are
>> +  removed.
>> +
>> +For sequential zone files, inconsistencies between an inode size and the amount
>> +of data writen in its zone, that is, the position of the file zone write
>> +pointer, can result from different events:
>> +* When the device write cache is enabled, a differed write error can occur
> 
> "a different write error"?

Nope. I really meant differed, as in "delayed" since the write command
succeeded but the cache flush for the data passed by the already completed
write command fails later. The sentence is not clear. I will clarify this
error pattern.

> 
>> +  resulting in the amount of data written in the zone being less than the inode
>> +  size.
> 
> Though I suspect that what you really mean to say is that errors can
> occur in ranges of previously completed writes can occur when the
> cache is flushed, hence less data being physically written than the
> OS has previously be told was written by the hardware. i.e. visible
> inode size goes backwards.

Yes, exactly. I will copy-paste this very clear explanation :)

>> +Finally, defective drives may change the condition of any zone to offline (zone
>> +dead) or read-only. Such changes, when discovered with the IO errors they can
>> +cause, are handled automatically regardless of the options specified at mount
>> +time. For offline zones, the action taken is similar to the action defined by
>> +the errors=zone-offline mount option. For read-only zones, the action used is
>> +as defined by the errors=zone-ro mount option.
> 
> Hmmmm. I think that's over-complicating things and takes control of
> error handling away from the user. That is, regardless of the reason
> for the error, if we get a write error and the user specified
> errors=remount-ro, the entire device should go read-only because
> that's what the user has told the filesystem to do on write error.

Yes, and that is the case. Any IO error with errors=remount-ro will turn
the FS read-only. What I tried to say here is that this option will not
affect the handling of zones that went offline (done by the device). For
these, the file will not even be read-only. The table will clarify that. I
also need to clarify the different causes for errors, e.g. "regular"
read-write errors due to bad sectors, excessive vibrations, etc, which are
generally recoverable (rewrite over bad sectors fixes the sector most of
the time) and the ones due to the device changing zones condition, which
are not recoverable (no condition can get these zones out of their bad state).

> This seems pretty user-unfriendly - giving them a way to control
> error handling behaviour and then ignoring it for specific errors
> despite the fact they mean exactly the same thing to the user: the
> write failed because a zone has gone bad since the last time that
> zone was accessed by the application....

I think it is only the explanation that is bad. The error control mount
options are enforced correctly as defined.

Thanks !

> 
> Cheers,
> 
> Dave.
>
diff mbox series

Patch

diff --git a/Documentation/filesystems/zonefs.txt b/Documentation/filesystems/zonefs.txt
new file mode 100644
index 000000000000..6f8b2004248f
--- /dev/null
+++ b/Documentation/filesystems/zonefs.txt
@@ -0,0 +1,310 @@ 
+ZoneFS - Zone filesystem for Zoned block devices
+
+Introduction
+============
+
+zonefs is a very simple file system exposing each zone of a zoned block device
+as a file. Unlike a regular POSIX-compliant file system with native zoned block
+device support (e.g. f2fs), zonefs does not hide the sequential write
+constraint of zoned block devices to the user. Files representing sequential
+write zones of the device must be written sequentially starting from the end
+of the file (append only writes).
+
+As such, zonefs is in essence closer to a raw block device access interface
+than to a full-featured POSIX file system. The goal of zonefs is to simplify
+the implementation of zoned block device support in applications by replacing
+raw block device file accesses with a richer file API, avoiding relying on
+direct block device file ioctls which may be more obscure to developers. One
+example of this approach is the implementation of LSM (log-structured merge)
+tree structures (such as used in RocksDB and LevelDB) on zoned block devices
+by allowing SSTables to be stored in a zone file similarly to a regular file
+system rather than as a range of sectors of the entire disk. The introduction
+of the higher level construct "one file is one zone" can help reducing the
+amount of changes needed in the application as well as introducing support for
+different application programming languages.
+
+Zoned block devices
+-------------------
+
+Zoned storage devices belong to a class of storage devices with an address
+space that is divided into zones. A zone is a group of consecutive LBAs and all
+zones are contiguous (there are no LBA gaps). Zones may have different types.
+* Conventional zones: there are no access constraints to LBAs belonging to
+  conventional zones. Any read or write access can be executed, similarly to a
+  regular block device.
+* Sequential zones: these zones accept random reads but must be written
+  sequentially. Each sequential zone has a write pointer maintained by the
+  device that keeps track of the mandatory start LBA position of the next write
+  to the device. As a result of this write constraint, LBAs in a sequential zone
+  cannot be overwritten. Sequential zones must first be erased using a special
+  command (zone reset) before rewriting.
+
+Zoned storage devices can be implemented using various recording and media
+technologies. The most common form of zoned storage today uses the SCSI Zoned
+Block Commands (ZBC) and Zoned ATA Commands (ZAC) interfaces on Shingled
+Magnetic Recording (SMR) HDDs.
+
+Solid State Disks (SSD) storage devices can also implement a zoned interface
+to, for instance, reduce internal write amplification due to garbage collection.
+The NVMe Zoned NameSpace (ZNS) is a technical proposal of the NVMe standard
+committee aiming at adding a zoned storage interface to the NVMe protocol.
+
+Zonefs Overview
+===============
+
+Zonefs exposes the zones of a zoned block device as files. The files
+representing zones are grouped by zone type, which are themselves represented
+by sub-directories. This file structure is built entirely using zone information
+provided by the device and so does not require any complex on-disk metadata
+structure.
+
+zonefs on-disk metadata
+-----------------------
+
+zonefs on-disk metadata is reduced to an immutable super block which
+persistently stores a magic number and optional feature flags and values. On
+mount, zonefs uses blkdev_report_zones() to obtain the device zone configuration
+and populates the mount point with a static file tree solely based on this
+information. File sizes come from the device zone type and write pointer
+position managed by the device itself.
+
+The super block is always written on disk at sector 0. The first zone of the
+device storing the super block is never exposed as a zone file by zonefs. If
+the zone containing the super block is a sequential zone, the mkzonefs format
+tool always "finishes" the zone, that is, it transitions the zone to a full
+state to make it read-only, preventing any data write.
+
+Zone type sub-directories
+-------------------------
+
+Files representing zones of the same type are grouped together under the same
+sub-directory automatically created on mount.
+
+For conventional zones, the sub-directory "cnv" is used. This directory is
+however created if and only if the device has usable conventional zones. If
+the device only has a single conventional zone at sector 0, the zone will not
+be exposed as a file as it will be used to store the zonefs super block. For
+such devices, the "cnv" sub-directory will not be created.
+
+For sequential write zones, the sub-directory "seq" is used.
+
+These two directories are the only directories that exist in zonefs. Users
+cannot create other directories and cannot rename nor delete the "cnv" and
+"seq" sub-directories.
+
+The size of the directories indicated by the st_size field of struct stat,
+obtained with the stat() or fstat() system calls, indicates the number of files
+existing under the directory.
+
+Zone files
+----------
+
+Zone files are named using the number of the zone they represent within the set
+of zones of a particular type. That is, both the "cnv" and "seq" directories
+contain files named "0", "1", "2", ... The file numbers also represent
+increasing zone start sector on the device.
+
+All read and write operations to zone files are not allowed beyond the file
+maximum size, that is, beyond the zone size. Any access exceeding the zone
+size is failed with the -EFBIG error.
+
+Creating, deleting, renaming or modifying any attribute of files and
+sub-directories is not allowed.
+
+The number of blocks of a file as reported by stat() and fstat() indicates the
+size of the file zone, or in other words, the maximum file size.
+
+Conventional zone files
+-----------------------
+
+The size of conventional zone files is fixed to the size of the zone they
+represent. Conventional zone files cannot be truncated.
+
+These files can be randomly read and written using any type of I/O operation:
+buffered I/Os, direct I/Os, memory mapped I/Os (mmap), etc. There are no I/O
+constraint for these files beyond the file size limit mentioned above.
+
+Sequential zone files
+---------------------
+
+The size of sequential zone files grouped in the "seq" sub-directory represents
+the file's zone write pointer position relative to the zone start sector.
+
+Sequential zone files can only be written sequentially, starting from the file
+end, that is, write operations can only be append writes. Zonefs makes no
+attempt at accepting random writes and will fail any write request that has a
+start offset not corresponding to the end of the file, or to the end of the last
+write issued and still in-flight (for asynchrnous I/O operations).
+
+Since dirty page writeback by the page cache does not guarantee a sequential
+write pattern, zonefs prevents buffered writes and writeable shared mappings
+on sequential files. Only direct I/O writes are accepted for these files.
+zonefs relies on the sequential delivery of write I/O requests to the device
+implemented by the block layer elevator. An elevator implementing the sequential
+write feature for zoned block device (ELEVATOR_F_ZBD_SEQ_WRITE elevator feature)
+must be used. This type of elevator (e.g. mq-deadline) is the set by default
+for zoned block devices on device initialization.
+
+There are no restrictions on the type of I/O used for read operations in
+sequential zone files. Buffered I/Os, direct I/Os and shared read mappings are
+all accepted.
+
+Truncating sequential zone files is allowed only down to 0, in which case, the
+zone is reset to rewind the file zone write pointer position to the start of
+the zone, or up to the zone size, in which case the file's zone is transitioned
+to the FULL state (finish zone operation).
+
+zonefs format options
+---------------------
+
+Several optional features of zonefs can be enabled at format time.
+* Conventional zone aggregation: ranges of contiguous conventional zones can be
+  aggregated into a single larger file instead of the default one file per zone.
+* File ownership: The owner UID and GID of zone files is by default 0 (root)
+  but can be changed to any valid UID/GID.
+* File access permissions: the default 640 access permissions can be changed.
+
+zonefs mount options
+--------------------
+
+zonefs defines several mount options allowing the user to control the file
+system behavior when write I/O errors occur and when inconsistencies between a
+file size and its zone write pointer position are discovered. The handling of
+read I/O errors is not changed by these options as long as no inode size
+corruption is detected.
+
+These options are as follows.
+* errors=remount-ro (default)
+  All write IO errors and errors due to a zone of the device going "bad"
+  (condition changed to offline or read-only), the file system is remounted
+  read-only after fixing the size and access permissions of the inode that
+  suffered the IO error.
+* errors=zone-ro
+  Any write IO error to a file zone result in the zone being considered as in a
+  read-only condition, preventing any further modification to the file. This
+  option does not affect the handling of errors due to offline zones. For these
+  zones, all accesses (read and write) are disabled.
+* errors=zone-offline
+  Any write IO error to a file zone result in the zone being considered as in
+  an offline condition. This implies that the file size is changed to 0 and all
+  read/write accesses to the file disabled, preventing all accesses by the user.
+* errors=repair
+  Any inconsistency between an inode size and its zone amount of written data
+  due to IO errors or external corruption are fixed without any change to file
+  access rights. This option does not affect the processing of zones that were
+  signaled as read-only or offline by the device. For read-only zones, the file
+  read accesses are disabled and for offline zones, all access permissions are
+  removed.
+
+For sequential zone files, inconsistencies between an inode size and the amount
+of data writen in its zone, that is, the position of the file zone write
+pointer, can result from different events:
+* When the device write cache is enabled, a differed write error can occur
+  resulting in the amount of data written in the zone being less than the inode
+  size.
+* Partial failures of large write I/O operations (e.g. one BIO of a multi-bio
+  large direct write fails) can result in the amount of data written in the
+  zone being larger than the inode size.
+* External action on the disk such as write, zone reset or zone finish
+  operations will change a file zone write pointer position resulting in a
+  reported amount of written data being different from the file inode size.
+
+Finally, defective drives may change the condition of any zone to offline (zone
+dead) or read-only. Such changes, when discovered with the IO errors they can
+cause, are handled automatically regardless of the options specified at mount
+time. For offline zones, the action taken is similar to the action defined by
+the errors=zone-offline mount option. For read-only zones, the action used is
+as defined by the errors=zone-ro mount option.
+
+Zonefs User Space Tools
+=======================
+
+The mkzonefs tool is used to format zoned block devices for use with zonefs.
+This tool is available on Github at:
+
+https://github.com/damien-lemoal/zonefs-tools
+
+zonefs-tools also includes a test suite which can be run against any zoned
+block device, including null_blk block device created with zoned mode.
+
+Examples
+--------
+
+The following formats a 15TB host-managed SMR HDD with 256 MB zones
+with the conventional zones aggregation feature enabled.
+
+# mkzonefs -o aggr_cnv /dev/sdX
+# mount -t zonefs /dev/sdX /mnt
+# ls -l /mnt/
+total 0
+dr-xr-xr-x 2 root root     1 Nov 25 13:23 cnv
+dr-xr-xr-x 2 root root 55356 Nov 25 13:23 seq
+
+The size of the zone files sub-directories indicate the number of files
+existing for each type of zones. In this example, there is only one
+conventional zone file (all conventional zones are aggregated under a single
+file).
+
+# ls -l /mnt/cnv
+total 137101312
+-rw-r----- 1 root root 140391743488 Nov 25 13:23 0
+
+This aggregated conventional zone file can be used as a regular file.
+
+# mkfs.ext4 /mnt/cnv/0
+# mount -o loop /mnt/cnv/0 /data
+
+The "seq" sub-directory grouping files for sequential write zones has in this
+example 55356 zones.
+
+# ls -lv /mnt/seq
+total 14511243264
+-rw-r----- 1 root root 0 Nov 25 13:23 0
+-rw-r----- 1 root root 0 Nov 25 13:23 1
+-rw-r----- 1 root root 0 Nov 25 13:23 2
+...
+-rw-r----- 1 root root 0 Nov 25 13:23 55354
+-rw-r----- 1 root root 0 Nov 25 13:23 55355
+
+For sequential write zone files, the file size changes as data is appended at
+the end of the file, similarly to any regular file system.
+
+# dd if=/dev/zero of=/mnt/seq/0 bs=4096 count=1 conv=notrunc oflag=direct
+1+0 records in
+1+0 records out
+4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00044121 s, 9.3 MB/s
+
+# ls -l /mnt/seq/0
+-rw-r----- 1 root root 4096 Nov 25 13:23 /mnt/seq/0
+
+The written file can be truncated to the zone size, preventing any further
+write operation.
+
+# truncate -s 268435456 /mnt/seq/0
+# ls -l /mnt/seq/0
+-rw-r----- 1 root root 268435456 Nov 25 13:49 /mnt/seq/0
+
+Truncation to 0 size allows freeing the file zone storage space and restart
+append-writes to the file.
+
+# truncate -s 0 /mnt/seq/0
+# ls -l /mnt/seq/0
+-rw-r----- 1 root root 0 Nov 25 13:49 /mnt/seq/0
+
+Since files are statically mapped to zones on the disk, the number of blocks of
+a file as reported by stat() and fstat() indicates the size of the file zone.
+
+# stat /mnt/seq/0
+  File: /mnt/seq/0
+  Size: 0         	Blocks: 524288     IO Block: 4096   regular empty file
+Device: 870h/2160d	Inode: 50431       Links: 1
+Access: (0640/-rw-r-----)  Uid: (    0/    root)   Gid: (    0/    root)
+Access: 2019-11-25 13:23:57.048971997 +0900
+Modify: 2019-11-25 13:52:25.553805765 +0900
+Change: 2019-11-25 13:52:25.553805765 +0900
+ Birth: -
+
+The number of blocks of the file ("Blocks") in units of 512B blocks gives the
+maximum file size of 524288 * 512 B = 256 MB, corresponding to the device zone
+size in this example. Of note is that the "IO block" field always indicates the
+minimum I/O size for writes and corresponds to the device physical sector size.
diff --git a/MAINTAINERS b/MAINTAINERS
index 089fd879632a..e9dcf8952573 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -18311,6 +18311,7 @@  L:	linux-fsdevel@vger.kernel.org
 T:	git git://git.kernel.org/pub/scm/linux/kernel/git/dlemoal/zonefs.git
 S:	Maintained
 F:	fs/zonefs/
+F:	Documentation/filesystems/zonefs.txt
 
 ZPOOL COMPRESSED PAGE STORAGE API
 M:	Dan Streetman <ddstreet@ieee.org>