Blobfuse2 is an open source project developed to provide a virtual filesystem backed by the Azure Storage. It uses the libfuse open source library (fuse3) to communicate with the Linux FUSE kernel module, and implements the filesystem operations using the Azure Storage REST APIs. This is the next generation blobfuse.
Blobfuse2 is stable and supported by Microsoft when used within its documented limits. Blobfuse2 supports high-performance reads and writes with strong consistency; however, it is recommended that multiple clients do not modify the same blob/file simultaneously to ensure data integrity. Blobfuse2 does not guarantee continuous synchronization of data written to the same blob/file using multiple clients or across multiple mounts of Blobfuse2 concurrently. If you modify an existing blob/file with another client while also reading that object, Blobfuse2 will not return the most up-to-date data. To ensure your reads see the newest blob/file data, disable all forms of caching at kernel (using direct-io
) as well as at Blobfuse2 level, and then re-open the blob/file.
Please submit an issue here for any issues/feature requests/questions.
This section will help you choose the correct config for Blobfuse2.
- Due to known data consistency issues when using Blobfuse2 in
block-cache
mode, it is strongly recommended that all Blobfuse2 installations be upgraded to version 2.3.2. For more information, see this. - As of version 2.3.0, blobfuse has updated its authentication methods. For Managed Identity, Object-ID based OAuth is solely accessible via CLI-based login, requiring Azure CLI on the system. For a dependency-free option, users may utilize Application/Client-ID or Resource ID based authentication.
streaming
mode is deprecated. Blobfuse2 will implicitly convert your streaming config to block-cache.
- Concurrent write operations on the same file using multiple handles is not checked for data consistency and may lead to incorrect data being written.
- A read operation on a file that is being written to simultaneously by another process or handle will not return the most up-to-date data.
- When copying files with trailing null bytes using
cp
utility to a Blobfuse2 mounted path, use--sparse=never
parameter to avoid data being trimmed. For example,cp --sparse=never src dest
. - In write operations, data written is persisted (or committed) to the Azure Storage container only when close, sync or flush operations are called by user application.
- Files cannot be modified if they were originally created with block-size different than the one configured.
- User applications must check the returned code (success/failure) for filesystem calls like read, write, close, flush, etc. If error is returned, the application must abort their respective operation.
- User applications must ensure that there is only one writer at a time for a given file.
- When dealing with very large files (in TiB), the block-size must be configured accordingly. Azure Storage supports only 50,000 blocks per blob.
This page lists various benchmarking results for HNS and FNS Storage account.
Visit this page to see list of supported linux distros.
- Mount an Azure storage blob container or datalake file system on Linux.
- Basic file system operations such as mkdir, opendir, readdir, rmdir, open, read, create, write, close, unlink, truncate, stat, rename
- Local caching to improve subsequent access times
- Block-Cache to support reading AND writing large files
- Parallel downloads and uploads to improve access time for large files
- Multiple mounts to the same container for read-only workloads
One of the biggest BlobFuse2 features is our brand new health monitor. It allows customers gain more insight into how their BlobFuse2 instance is behaving with the rest of their machine. Visit here to set it up.
- Blobfuse2 is fuse3 compatible (other than Ubuntu-18 and Debian-9, where it still runs with fuse2)
- Support for higher service version offering latest and greatest of azure storage features (supported by azure go-sdk)
- Set blob tier while uploading the data to storage
- Attribute cache invalidation based on timeout
- For flat namespace accounts, user can configure default permissions for files and folders
- Improved cache eviction algorithm for file cache to control disk footprint of blobfuse2
- Improved cache eviction algorithm for streamed buffers to control memory footprint of blobfuse2
- Utility to convert blobfuse CLI and config parameters to a blobfuse2 compatible config for easy migration
- CLI to mount Blobfuse2 with legacy Blobfuse config and CLI parameters (Refer to Migration guide for this)
- Version check and upgrade prompting
- Option to mount a sub-directory from a container
- CLI to mount all containers (with a allowlist and denylist) in a given storage account
- CLI to list all blobfuse2 mount points
- CLI to unmount one, multiple or all blobfuse2 mountpoints
- Option to dump logs to syslog or a file on disk
- Support for config file encryption and mounting with an encrypted config file via a passphrase (CLI or environment variable) to decrypt the config file
- CLI to check or update a parameter in the encrypted config
- Set MD5 sum of a blob while uploading
- Validate MD5 sum on download and fail file open on mismatch
- Large file writing through write Block-Cache
- 'git clone' operation is 25% faster (tested with vscode repo cloning)
- ResNet50 image classification job is 7-8% faster (tested with 1.3 million images)
- Regular file uploads are 10% faster
- Verified listing of 1-Billion files in a directory (which v1.x does not support)
You can install Blobfuse2 by cloning this repository. In the workspace root execute below commands to build the binary.
- sudo apt install fuse3 libfuse3-dev gcc
- go build -o blobfuse2
The general format of the Blobfuse2 commands is blobfuse2 [command] [arguments] --[flag-name]=[flag-value]
help
- Help about any commandmount
- Mounts an Azure container as a filesystem. The supported containers include- Azure Blob Container
- Azure Datalake Gen2 Container
mount all
- Mounts all the containers in an Azure account as a filesystem. The supported storage services includemount list
- Lists all Blobfuse2 filesystems.secure decrypt
- Decrypts a config file.secure encrypt
- Encrypts a config file.secure get
- Gets value of a config parameter from an encrypted config file.secure set
- Updates value of a config parameter.unmount
- Unmounts the Blobfuse2 filesystem.unmount all
- Unmounts all Blobfuse2 filesystems.
To see a list of commands, type blobfuse2 -h
and then press the ENTER key.
To learn about a specific command, just include the name of the command (For example: blobfuse2 mount -h
).
- Mount with blobfuse2
- blobfuse2 mount --config-file=
- Mount blobfuse2 using legacy blobfuse config and cli parameters
- blobfuse2 mountv1
- Mount all containers in your storage account
- blobfuse2 mount all --config-file=
- List all mount instances of blobfuse2
- blobfuse2 mount list
- Unmount blobfuse2
- sudo fusermount3 -u
- Unmount all blobfuse2 instances
- blobfuse2 unmount all
- Note: Blobfuse2 accepts all CLI parameters that Blobfuse does, but may ignore parameters that are no longer applicable.
- General options
--config-file=<PATH>
: The path to the config file.--log-level=<LOG_*>
: The level of logs to capture.--log-file-path=<PATH>
: The path for the log file.--foreground=true
: Mounts the system in foreground mode.--read-only=true
: Mount container in read-only mode.--default-working-dir
: The default working directory to store log files and other blobfuse2 related information.--disable-version-check=true
: Disable the blobfuse2 version check.--secure-config=true
: Config file is encrypted suing 'blobfuse2 secure` command.--passphrase=<STRING>
: Passphrase used to encrypt/decrypt config file.--wait-for-mount=<TIMEOUT IN SECONDS>
: Let parent process wait for given timeout before exit to ensure child has started.--block-cache
: To enable block-cache instead of file-cache. This works only when mounted without any config file.--lazy-write
: To enable async close file handle call and schedule the upload in background.
- Attribute cache options
--attr-cache-timeout=<TIMEOUT IN SECONDS>
: The timeout for the attribute cache entries.--no-symlinks=true
: To improve performance disable symlink support.
- Storage options
--container-name=<CONTAINER NAME>
: The container to mount.--cancel-list-on-mount-seconds=<TIMEOUT IN SECONDS>
: Time for which list calls will be blocked after mount. ( prevent billing charges on mounting)--virtual-directory=true
: Support virtual directories without existence of a special marker blob for block blob account.--subdirectory=<path>
: Subdirectory to mount instead of entire container.--disable-compression:false
: Disable content encoding negotiation with server. If blobs have 'content-encoding' set to 'gzip' then turn on this flag.--use-adls=false
: Specify configured storage account is HNS enabled or not. This must be turned on when HNS enabled account is mounted.--cpk-enabled=true
: Allows mounting containers with cpk. Use config file or env variables to set cpk encryption key and cpk encryption key sha.
- File cache options
--file-cache-timeout=<TIMEOUT IN SECONDS>
: Timeout for which file is cached on local system.--tmp-path=<PATH>
: The path to the file cache.--cache-size-mb=<SIZE IN MB>
: Amount of disk cache that can be used by blobfuse. Default - 80% of free disk space.--high-disk-threshold=<PERCENTAGE>
: If local cache usage exceeds this, start early eviction of files from cache.--low-disk-threshold=<PERCENTAGE>
: If local cache usage comes below this threshold then stop early eviction.--sync-to-flush=false
: Sync call will force upload a file to storage container if this is set to true, otherwise it just evicts file from local cache.
- Block-Cache options
--block-cache-block-size=<SIZE IN MB>
: Size of a block to be downloaded as a unit.--block-cache-pool-size=<SIZE IN MB>
: Size of pool to be used for caching. This limits total memory used by block-cache. Default - 80% of free memory available.--block-cache-path=<PATH>
: Path where downloaded blocks will be persisted. Not providing this parameter will disable the disk caching.--block-cache-disk-size=<SIZE IN MB>
: Disk space to be used for caching. Default - 80% of free disk space.--block-cache-disk-timeout=<seconds>
: Timeout for which disk cache is valid.--block-cache-prefetch=<Number of blocks>
: Number of blocks to prefetch at max when sequential reads are in progress. Default - 2 times number of CPU cores.--block-cache-parallelism=<count>
: Number of parallel threads doing upload/download operation. Default - 3 times number of CPU cores.--block-cache-prefetch-on-open=true
: Start prefetching on open system call instead of waiting for first read. Enhances perf if file is read sequentially from offset 0.
- Fuse options
--attr-timeout=<TIMEOUT IN SECONDS>
: Time the kernel can cache inode attributes.--entry-timeout=<TIMEOUT IN SECONDS>
: Time the kernel can cache directory listing.--negative-timeout=<TIMEOUT IN SECONDS>
: Time the kernel can cache non-existance of file or directory.--allow-other
: Allow other users to have access this mount point.--disable-writeback-cache=true
: Disallow libfuse to buffer write requests if you must strictly open files in O_WRONLY or O_APPEND mode.--ignore-open-flags=true
: Ignore the append and write only flag since O_APPEND and O_WRONLY is not supported with writeback caching.
- General options
AZURE_STORAGE_ACCOUNT
: Specifies the storage account to be connected.AZURE_STORAGE_ACCOUNT_TYPE
: Specifies the account type 'block' or 'adls'AZURE_STORAGE_ACCOUNT_CONTAINER
: Specifies the name of the container to be mountedAZURE_STORAGE_BLOB_ENDPOINT
: Specifies the blob endpoint to use. Defaults to *.blob.core.windows.net, but is useful for targeting storage emulators.AZURE_STORAGE_AUTH_TYPE
: Overrides the currently specified auth type. Case insensitive. Options: Key, SAS, MSI, SPN
- Account key auth:
AZURE_STORAGE_ACCESS_KEY
: Specifies the storage account key to use for authentication.
- SAS token auth:
AZURE_STORAGE_SAS_TOKEN
: Specifies the SAS token to use for authentication.
- Managed Identity auth:
AZURE_STORAGE_IDENTITY_CLIENT_ID
: Only one of these three parameters are needed if multiple identities are present on the system.AZURE_STORAGE_IDENTITY_OBJECT_ID
: Only one of these three parameters are needed if multiple identities are present on the system.AZURE_STORAGE_IDENTITY_RESOURCE_ID
: Only one of these three parameters are needed if multiple identities are present on the system.MSI_ENDPOINT
: Specifies a custom managed identity endpoint, as IMDS may not be available under some scenarios. Uses theMSI_SECRET
parameter as theSecret
header.MSI_SECRET
: Specifies a custom secret for an alternate managed identity endpoint.
- Service Principal Name auth:
AZURE_STORAGE_SPN_CLIENT_ID
: Specifies the client ID for your application registrationAZURE_STORAGE_SPN_TENANT_ID
: Specifies the tenant ID for your application registrationAZURE_STORAGE_AAD_ENDPOINT
: Specifies a custom AAD endpoint to authenticate againstAZURE_STORAGE_SPN_CLIENT_SECRET
: Specifies the client secret for your application registration.AZURE_STORAGE_AUTH_RESOURCE
: Scope to be used while requesting for token.
- Proxy Server:
http_proxy
: The proxy server address. Example:10.1.22.4:8080
.https_proxy
: The proxy server address when https is turned off forcing http. Example:10.1.22.4:8080
.
- CPK options:
AZURE_STORAGE_CPK_ENCRYPTION_KEY
: Customer provided base64-encoded AES-256 encryption key value.AZURE_STORAGE_CPK_ENCRYPTION_KEY_SHA256
: Base64-encoded SHA256 of the cpk encryption key.
- Custom component options:
BLOBFUSE_PLUGIN_PATH
: Specifies plugin file path as a colon-separated list of.so
files. Example BLOBFUSE_PLUGIN_PATH="/path/to/plugin1.so:/path/to/plugin2.so".
Below diagrams guide you to choose right configuration for your workloads.
- Choose right Auth mode
- Choose right caching for Read-Only workloads
- Choose right caching for Read-Write workloads
- Choose right block-cache configuration
- Choose right file-cache configuration
- Sample File Cache Config
- Sample Block-Cache Config
- All Config options
- How do I generate a SAS with permissions for rename? az cli has a command to generate a sas token. Open a command prompt and make sure you are logged in to az cli. Run the following command and the sas token will be displayed in the command prompt. az storage container generate-sas --account-name --account-key -n --permissions dlrwac --start <today's date ex: 2021-03-26> --expiry
- Why do I get EINVAL on opening a file with WRONLY or APPEND flags? To improve performance, Blobfuse2 by default enables writeback caching, which can produce unexpected behavior for files opened with WRONLY or APPEND flags, so Blobfuse2 returns EINVAL on open of a file with those flags. Either use disable-writeback-caching to turn off writeback caching (can potentially result in degraded performance) or ignore-open-flags (replace WRONLY with RDWR and ignore APPEND) based on your workload.
- How to mount blobfuse2 inside a container? Refer to 'docker' folder in this repo. It contains a sample 'Dockerfile'. If you wish to create your own container image, try 'buildandruncontainer.sh' script, it will create a container image and launch the container using current environment variables holding your storage account credentials.
- Why am I not able to see the updated contents of file(s), which were updated through means other than Blobfuse2 mount?
If your use-case involves updating/uploading file(s) through other means and you wish to see the updated contents on Blobfuse2 mount then you need to disable kernel page-cache.
-o direct_io
CLI parameter is the option you need to use while mounting. Along with this, setfile-cache-timeout=0
and all other libfuse caching parameters should also be set to 0. User shall be aware that disabling kernel cache can result into more calls to Azure Storage which will have cost and performance implications.
- mkfifo : fifo creation is not supported by blobfuse2 and this will result in "function not implemented" error
- chown : Change of ownership is not supported by Azure Storage hence Blobfuse2 does not support this.
- Creation of device files or pipes is not supported by Blobfuse2.
- Blobfuse2 does not support extended-attributes (x-attrs) operations
- Blobfuse2 does not support lseek() operation on directory handles. No error is thrown but it will not work as expected.
-
Blobfuse2 does not support overlapping mount paths. While running multiple instances of Blobfuse2 make sure each instance has a unique and non-overlapping mount point.
-
Blobfuse2 does not support co-existance with NFS on same mount path. Behaviour in this case is undefined.
-
For block blob accounts, where data is uploaded through other means, Blobfuse2 expects special directory marker files to exist in container. In absence of this few file operations might not work. For e.g. if you have a blob 'A/B/c.txt' then special marker files shall exists for 'A' and 'A/B', otherwise opening of 'A/B/c.txt' will fail. Once a 'ls' operation is done on these directories 'A' and 'A/B' you will be able to open 'A/B/c.txt' as well. Possible workaround to resolve this from your container is to either
create the directory marker files manually through portal or run 'mkdir' command for 'A' and 'A/B' from blobfuse. Refer me for details on this.
-
In case of BlockBlob accounts, ACLs are not supported by Azure Storage so Blobfuse2 will by default return success for 'chmod' operation. However it will work fine for Gen2 (DataLake) accounts.
-
When Blobfuse2 is mounted on a container, SYS_ADMIN privileges are required for it to interact with the fuse driver. If container is created without the privilege, mount will fail. Sample command to spawn a docker container is
docker run -it --rm --cap-add=SYS_ADMIN --device=/dev/fuse --security-opt apparmor:unconfined <environment variables> <docker image>
-
In case of
mount all
system may limit on number of containers you can mount in parallel (when you go above 100 containers). To increase this system limit use below commandecho 256 | sudo tee /proc/sys/fs/inotify/max_user_instances
-
Refer this for block-cache limitations.
By default, Blobfuse2 will log to syslog. The default settings will, in some cases, log relevant file paths to syslog. If this is sensitive information, turn off logging or set log-level to LOG_ERR.
This project is licensed under MIT.
This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.microsoft.com.
When you submit a pull request, a CLA-bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., label, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.
This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact [email protected] with any additional questions or comments.