F2FS

From Wikipedia, the free encyclopedia
Jump to: navigation, search
F2FS
Developer(s) Samsung Electronics, Motorola Mobility, and Huawei
Full name Flash-Friendly File System
Introduced v3.8, 2012-12-20[1] with Linux
Structures
Directory contents multi-level hash table
File allocation bitmap (free space), table
Limits
Max. volume size 16 TB
Max. file size 3.94 TB
Max. number of files Depends on volume size
Max. filename length 255 bytes
Features
Dates recorded modification (mtime), attribute modification (ctime), access (atime)
Date resolution 1 ns
Attributes POSIX, extended attributes
File system permissions POSIX, ACL
Transparent compression No
Transparent encryption Yes
Other
Supported operating systems Linux
Website f2fs.wiki.kernel.org

F2FS (Flash-Friendly File System) is a flash file system initially developed by Samsung Electronics for the Linux kernel.[2]

The motive for F2FS was to build a file system that, from the start, takes into account the characteristics of NAND flash memory-based storage devices (such as solid-state disks, eMMC, and SD cards), which are widely used in computer systems ranging from mobile devices to servers.

F2FS was designed on a basis of a log-structured file system approach, which is adapted to newer forms of storage. Jaegeuk Kim, the principal F2FS author, has stated that it remedies some known issues[2] of the older log-structured file systems, such as the snowball effect of wandering trees and high cleaning overhead. In addition, since a NAND-based storage device shows different characteristics according to its internal geometry or flash memory management scheme (such as the Flash Translation Layer or FTL), it supports various parameters not only for configuring on-disk layout, but also for selecting allocation and cleaning algorithms.

Features[edit]

Design[edit]

On-disk layout[edit]

F2FS divides the whole volume into a number of segments, each of which is fixed at 2 MB. A section is composed of consecutive segments, and a zone consists of a set of sections. By default, section and zone sizes are set to the same size, but users can easily modify the size with mkfs.

F2FS splits the entire volume into six areas, and all except the superblock area consist of multiple segments as described below.

Superblock (SB)
The SB is located at the beginning of the partition. There are two copies to avoid file-system corruption. It contains basic partition information and some default F2FS parameters.
Checkpoint (CP)
The CP contains file system information, bitmaps for valid NAT/SIT sets, orphan inode lists, and summary entries of current active segments.
Segment Information Table (SIT)
The SIT contains the valid block count and validity bitmap of all the Main Area blocks.
Node Address Table (NAT)
The NAT is an address table for the Main Area node blocks.
Segment Summary Area (SSA)
The SSA contains entries which contain the owner information of the Main Area data and node blocks.
Main Area
The main area contains file and directory data and their indices.

In order to avoid misalignment between file system and flash storage, F2FS aligns the start block address of the CP with the segment size. It also aligns the Main Area start block address with the zone size by reserving some segments in the SSA area.

Metadata structure[edit]

F2FS uses the checkpoint scheme to maintain file system integrity. At mount time, F2FS first tries to find the last valid checkpoint data by scanning the CP area. In order to reduce the scanning time, F2FS uses only two copies of the CP. One of them always indicates the last valid data, which is called a shadow copy mechanism. In addition to the CP, the NAT and SIT also use the shadow copy mechanism. For file system consistency, each CP points to which NAT and SIT copies are valid.

Index structure[edit]

The key data structure is the "node". Similar to traditional file structures, F2FS has three types of nodes: inode, direct node, indirect node. F2FS assigns 4 KB to an inode block which contains 923 data block indices, two direct node pointers, two indirect node pointers, and one double indirect node pointer as described below. A direct node block contains 1018 data block indices, and an indirect node block contains 1018 node block indices. Thus, one inode block (i.e., a file) covers:

4 KB × (923 + 2×1018 + 2×10182 + 10183) = 3.94 TB

Note that all the node blocks are mapped by the NAT, which means that the location of each node is translated by the NAT. To mitigate the wandering tree problem, F2FS is able to cut off the propagation of node updates caused by leaf data writes.

Directory structure[edit]

A directory entry (dentry) occupies 11 bytes, which consists of the following attributes.

A directory entry structure
hash hash value of the file name
ino inode number
len the length of file name
type file type such as directory, symlink, etc.

A dentry block consists of 214 dentry slots and file names. A bitmap is used to represent whether each dentry is valid or not. A dentry block occupies 4 KB and has the following composition:

Dentry Block (4 K) = bitmap (27 bytes) + reserved (3 bytes) +
                      dentries (11 * 214 bytes) + file name (8 * 214 bytes)

F2FS implements multi-level hash tables for the directory structure. Each level has a hash table with a dedicated number of hash buckets as shown below. Note that "A(2B)" means a bucket includes 2 data blocks.

Term
A indicates bucket
B indicates block
N indicates MAX_DIR_HASH_DEPTH
level #0    A(2B)
level #1    A(2B) - A(2B)
level #2    A(2B) - A(2B) - A(2B) - A(2B)
    ...
level #N/2  A(2B) - A(2B) - A(2B) - A(2B) - A(2B) - ... - A(2B)
    ...
level #N    A(4B) - A(4B) - A(4B) - A(4B) - A(4B) - ... - A(4B)

When F2FS finds a file name in a directory, first a hash value of the file name is calculated. Then, F2FS scans the hash table in level #0 to find the dentry consisting of the file name and its inode number. If not found, F2FS scans the next hash table in level #1. In this way, F2FS scans hash tables in each levels incrementally from 1 to N. In each level F2FS needs to scan only one bucket determined by the following equation, which shows O(log(# of files)) complexity.

 bucket number to scan in level #n = (hash value) % (# of buckets in level #n)

In the case of file creation, F2FS finds empty consecutive slots that cover the file name. F2FS searches the empty slots in the hash tables of whole levels from 1 to N in the same way as the lookup operation.

Default block allocation[edit]

At runtime, F2FS manages six active logs inside the "Main Area:" Hot/Warm/Cold node and Hot/Warm/Cold data.

Block allocation policy
Hot node Contains direct node blocks of directories.
Warm node Contains direct node blocks except hot node blocks.
Cold node Contains indirect node blocks.
Hot data Contains dentry blocks.
Warm data Contains data blocks except hot and cold data blocks.
Cold data Contains multimedia data or migrated data blocks.

LFS has two schemes for free space management: threaded log and copy-and-compaction. The copy-and-compaction scheme which is known as cleaning, is well-suited for devices showing very good sequential write performance, since free segments are served all the time for writing new data. However, it suffers from cleaning overhead during high utilization. Conversely, the threaded log scheme suffers from random writes, but no cleaning process is needed. F2FS adopts a hybrid scheme where the copy-and-compaction scheme is adopted by default, but the policy is dynamically changed to the threaded log scheme according to the file system status.

In order to align F2FS with underlying flash-based storage, F2FS allocates a segment in a unit of a section. F2FS expects the section size to be the same as the garbage collection unit size in FTL. With respect to the mapping granularity in FTL, F2FS allocates each section of the active logs to as many different zones as possible. FTL can write the active log data into one allocation unit according to its mapping granularity.

Cleaning process[edit]

F2FS does cleaning both on demand, and in the background. On-demand cleaning is triggered when there are not enough free segments to serve VFS calls. The background cleaner is executed by a kernel thread, and triggers the cleaning job when the system is idle.

F2FS supports two victim selection policies: greedy, and cost-benefit algorithms. In the greedy algorithm, F2FS selects a victim segment having the smallest number of valid blocks. In the cost-benefit algorithm, F2FS selects a victim segment according to the segment age and the number of valid blocks in order to address the log block thrashing problem present in the greedy algorithm. F2FS uses the greedy algorithm for on-demand cleaning, the background cleaner uses the cost-benefit algorithm.

In order to identify whether the data in the victim segment are valid or not, F2FS manages a bitmap. Each bit represents the validity of a block, and the bitmap is composed of a bit stream covering whole blocks in the Main Area.

See also[edit]

References[edit]

  1. ^ Michael Larabel (2012-12-22). "F2FS File-System Merged Into Linux 3.8 Kernel". Phoronix. Retrieved 2016-05-25. 
  2. ^ a b Jaegeuk Kim (2012-10-05). "f2fs: introduce flash-friendly file system". Retrieved 2016-05-25. 
  3. ^ Jaegeuk Kim (2014-09-22). "f2fs: introduce FITRIM in f2fs_ioctl". 
  4. ^ Chao Yu (2015-10-26). "f2fs: support file defragment". 
  5. ^ Jaegeuk Kim (2013-08-26). "f2fs: add flags for inline xattrs". 
  6. ^ Huajun Li (2013-11-10). "f2fs: Enable f2fs support inline data". 
  7. ^ Chao Yu (2014-09-24). "f2fs: support inline dir". 
  8. ^ Jaegeuk Kim (2014-09-20). "f2fs-tools: release 1.4.0". 
  9. ^ Jaegeuk Kim (2014-09-25). "f2fs: support atomic_write feature for database". 
  10. ^ Jaegeuk Kim (2015-06-24). "f2fs updates for v4.2". 
  11. ^ Jaegeuk Kim (2016-04-25). "resize.f2fs: support to expand partition size". 
  12. ^ Chao Yu (2015-12-17). "f2fs: support data flush in background". 
  13. ^ Chao Yu (2015-01-25). "f2fs: enable rb-tree extent cache". 

External links[edit]