File System Considerations

This chapter is intended to help you decide which file system is most suitable for your application. It points out specific features and properties of the file systems supported by the ESP-IDF, which are important in typical use-cases rather than describing all the specifics or comparing implementation details. Technical details for each file system are available in their corresponding documentation.

Currently, there are 3 file systems supported by the ESP-IDF framework:

All of them are based on 3rd-party libraries connected to the IDF through various wrappers and modifications.

NVS library (Non-Volatile Storage) is involved in the considerations too, since it can be seen as a type of file system providing specific behaviour not available in the other implementations, though NVS' primary goal is different. ESP-IDF provides comfort APIs to handle mounting and dismounting the file systems in unified way, file and directory access is implemented via C/POSIX standard file APIs. Thus, all the applications can use the same interface regardless of specific underlying system. The only exception is the NVS, whose details explained further in this document.

The most significant properties and features of above-mentioned file systems are summarised in the following table:

FatFS

SPIFFS

LittleFS

Features

  • Implements MS FAT12, FAT16, FAT32 and optionally exFAT variants

  • General purpose filesystem, widely compatible across most HW platforms

  • Well documented

  • Thread safe

  • Developed for NOR flash devices on embedded systems, low RAM usage

  • Implements static wear levelling

  • Limited documentation, no ongoing development

  • Thread safe

  • Designed as fail-safe, with own wear levelling and with fixed amount of RAM usage independent on the file system size

  • Well documented

  • Thread safe

Storage units and limits

  • Clusters (1-128 sectors)

  • Supported sector sizes: 512 B, 4096 B

  • FAT12: cluster size 512 B - 8 kB, max 4085 clusters

  • FAT16: cluster size 512 B - 64 kB, max 65525 clusters

  • FAT32: cluster size 512 B - 32 kB, max 268435455 clusters

  • Logical pages, logical blocks (consists of pages)

  • Typical setup: page = 256 B, block = 64 kB

  • Blocks, metadata pairs

  • Typical block size: 4 kB

Wear Levelling

Optional (for SPI Flash)

Integrated

Integrated

Minimum partition size

  • 128 sectors With wear levelling on (WL sector=4096B):

  • plus 4 sectors at least

  • real number given by WL configuration (Safe, Perf)

  • 6 logical blocks

  • 8 pages per block

Not specified, theoretically 2 blocks

Maximum partition size

  • FAT12: approx. 32 MB with 8 kB clusters

  • FAT16: approx. 4 GB with 64 kB clusters (theoretical)

  • FAT32: approx. 8 TB with 32 kB clusters (theoretical)

Absolute maximum not specified More than 1024 pages per block not recommended

Not specified, theoretically around 2 GB

Directory Support

  • Yes (max 65536 entries in a common FAT directory)

  • Limitations:
    • FAT12: max 224 files in the Root directory

    • FAT16: max 512 files in the Root directory

    • FAT32: the Root is just another directory

No

Yes

Power failure protection

No

Partial (see the SPIFFS notes)

Yes (integrated)

Encryption support

Yes

No

Yes

Supported targets

  • SPI Flash (NOR)

  • SD cards

SPI Flash (NOR)

  • SPI Flash (NOR)

  • SD cards (IDF >= v5.0)

For file systems performance comparison using various configurations and parameters, see Storage performance benchmark example storage/perf_benchmark.

FatFS

The most supported file system, recommended for common applications - file/directory operations, data storage, logging, etc. Automatic resolution of specific FAT system type, widely compatible with PC or other platforms. Supports partition encryption, read-only mode, optional wear-levelling for SPI Flash (SD cards use own built-in WL), equipped with auxiliary host side tools (generators and parsers, Python scripts). Supports SDMMC access. The biggest weakness is its low resilience against sudden power-off events. To mitigate such a scenario impact, the IDF FatFS default setup deploys 2 FAT table copies. This option can be disabled by setting esp_vfs_fat_mount_config_t::use_one_fat flag (the 2-FAT processing is fully handled by the FatFS library). See also related examples.

Related documents:

Examples:

SPIFFS

SPIFFS is a file system providing certain level of power-off safety (see repair-after-restart function esp_spiffs_check()) and built-in wear levelling. It tend to become slow down when exceeding around 70% of dedicated partition size due to its garbage collector implementation, and it also doesn't support directories. It is useful for applications depending only on few files (possibly large) and requiring high level of consistency. Generally, the SPIFFS needs less RAM resources than FatFS and supports flash chips up to 128MB in size. Please keep in mind the SPIFFS is not being developed and maintained anymore, so consider precisely whether its advantages for your project really prevail over the other file systems.

Related documents:

Examples:

LittleFS

LittleFS is a block based file system designed for microcontrollers and embedded devices. It provides a good level of power failure resilience, implements dynamic wear levelling and has very low RAM requirements, the system has configurable limits and integrated SD/MMC card support. It is a recommended choice for general type of application, the only disadvantage is the file system not being natively compatible with other platforms (unlike FAT).

LittleFS is available as external component in the ESP Registry, see LittleFS component page for the details on including the file system into your project.

Related documents:

Examples:

NVS Library

Non-volatile Storage (NVS) is useful for applications depending on handling numerous key-value pairs, for instance industrial configuration. Original goal of the NVS library is exactly to store the system configuration, thus there were no assupmtions about partitions larger than 128kB. However, over time the library has proven to be power-off resilient up to some extent and has been used many times as a regular file system. It is generally not recommended approach, still you are free to go this way as long as you know what to expect. As any other file system, even the NVS has certain overhead in terms of extra disk space and RAM use. NVS also supports multithreading safety, except for its initialisation and shutdown functions.

Points to keep in mind when developing NVS related code:

  • NVS partition must have 4 pages at least, 1 page must stay free to keep the NVS page management working.

  • NVS doesn't implement any sort of wear levelling function, as it is designed to minimise erasing of underlying flash device (value updates done by appending, page rotation). Still, it is relatively easy to exhaust your flash chip lifetime once your application makes frequent updates to the NVS store, especially on small partitions.

  • Be careful about using NVS blobs: internal overhead for handling this data type is quite high - for instance, the first page must have minimum 4 free entries to process any blob update, and the blob cost is 2 entries in this case. Not meeting the conditions causes ESP_ERR_NVS_NOT_ENOUGH_SPACE error (0x00001105). Blob storage on other pages consumes 1 entry plus the data space, still it can cause fast rotation over the whole NVS space, resulting in out-of-memory type of errors.

  • Do not use blob size too close to the NVS page size (4kB). Such an entry actually consumes 2 pages, and the NVS partition space is typically exhausted much sooner than expected.

  • NVS is not suitable for logging or similar use cases.

Further details available at NVS documentation page (Non-Volatile Storage).

Related documents:

Examples:

File handling design considerations

Here are several recommendation for building reliable storage features into your application:

  • Use C Standard Library file APIs (ISO or POSIX) wherever possible. This high-level interface guarantees you will not need to change much, if it comes for instance to switching to a different file system. All the IDF supported file systems work as underlying layer for C STDLIB calls, so the specific file system details are nearly transparent to the application code. The only parts unique to each single system are formatting, mounting and diagnostic/repair functions

  • Keep the file system dependent code separated, use wrappers to allow minimum change updates

  • Design reasonable structure of your application file storage:
    • Distribute the load evenly, if possible. Use meaningful number of directories/subdirectories (for instance FAT12 can keep only 224 record in its root directory).

    • Avoid using too many files or too large files (though the latter usually causes less troubles than the former). Each file equals to a record in the system's internal "database", which can easily end up in the necessary overhead consuming more space than the data stored. Even worse case is exhausting the filesystem's resources and subsequent failure of the application - which can happen really quickly in embedded systems' environment.

    • Be cautious about number of write or erase operations performed in SPI Flash memory (for example, each write in the FatFS involves full erase of the area to be written). NOR Flash devices typically survive 100.000+ erase cycles per sector, and their lifetime is extended by the Wear-Levelling mechanism (implemented as a standalone component in corresponding driver stack, transparent from the application's perspective). The Wear-Levelling algorithm rotates the Flash memory sectors all around given partition space, so it requires some disk space available for the virtual sector shuffle. If you create "well-tailored" partition with the minimum space needed and manage to fill it with your application data, the Wear Levelling becomes ineffective and your device would degrade quickly. Projects with Flash write frequency around 500ms are fully capable to destroy average ESP32 flash in few days time (real world example).

    • With the previous point given, consider using reasonably large partitions to ensure safe margins for your data. It is usually cheaper to invest into extra Flash space than to forcibly resolve troubles unexpectedly happening in the field.

    • Think twice before deciding for specific file system - they are not 100% equal and each application has own strategy and requirements. For instance, the NVS is not suitable for storing a production data, as its design doesn't deal well with too many items being stored (recommended maximum for NVS partition size would be around 128kB).

Encrypting partitions

ESP32-S2 based chips provide several features to encrypt the contents of various partitions within chip's main SPI flash memory. All the necessary information can be found in chapters Flash Encryption and NVS Encryption. Both variants use the AES family of algorithms, the Flash Encryption provides hardware-driven encryption scheme and is transparent from the software's perspective, whilst the NVS Encryption is a software feature implemented using mbedTLS component (though the mbedTLS can internally use the AES hardware accelerator, if available on given chip model). The latter requires the Flash Encryption enabled as the NVS Encryption needs a proprietary encrypted partition to hold its keys, and the NVS internal structure is not compatible with the Flash Encryption design. Therefore, both features come separate.

Given storage security scheme and the ESP32-S2 chips design result into a few implications which may not be fully obvious in the main documents:

  • The Flash encryption applies only to the main SPI Flash memory, due to its cache module design (all the "transparent" encryption APIs run over this cache). This implies that external flash partitions cannot be encrypted using the native Flash Encryption means.

  • External partition encryption can be deployed by implementing custom encrypt/decrypt code in appropriate driver APIs - either by implementing own SPI flash driver (see storage/custom_flash_driver) or by customising higher levels in the driver stack, for instance by providing own FatFS disk IO layer.


Was this page helpful?