Blobfuse2 is an open source project developed to provide a virtual filesystem backed by the Azure Storage. It uses the libfuse open source library (fuse3) to communicate with the Linux FUSE kernel module, and implements the filesystem operations using the Azure Storage REST APIs. This is the next generation blobfuse
Blobfuse2 is stable, and is supported by Microsoft provided that it is used within its limits documented here. Blobfuse2 supports both reads and writes however, it does not guarantee continuous sync of data written to storage using other APIs or other mounts of Blobfuse2. For data integrity it is recommended that multiple sources do not modify the same blob/file. Please submit an issue here for any issues/feature requests/questions.
This section will help you choose the correct config for Blobfuse2.
- We have seen some customer issues around files getting corrupted when
streaming
is used in write mode. Kindly avoid using this feature for write while we investigate and resolve it. - You can now use block-cache instead of streaming for both read and write workflows, which offers much better performance compared to streaming. To enable
block-cache
instead ofstreaming
, use--block-cache
in CLI param orblock-cache
as component in config file instead ofstreaming
.
Visit this page to see list of supported linux distros.
- Mount an Azure storage blob container or datalake file system on Linux.
- Basic file system operations such as mkdir, opendir, readdir, rmdir, open, read, create, write, close, unlink, truncate, stat, rename
- Local caching to improve subsequent access times
- Streaming/Block-Cache to support reading AND writing large files
- Parallel downloads and uploads to improve access time for large files
- Multiple mounts to the same container for read-only workloads
One of the biggest BlobFuse2 features is our brand new health monitor. It allows customers gain more insight into how their BlobFuse2 instance is behaving with the rest of their machine. Visit here to set it up.
- Blobfuse2 is fuse3 compatible (other than Ubuntu-18 and Debian-9, where it still runs with fuse2)
- Support for higher service version offering latest and greatest of azure storage features (supported by azure go-sdk)
- Set blob tier while uploading the data to storage
- Attribute cache invalidation based on timeout
- For flat namespace accounts, user can configure default permissions for files and folders
- Improved cache eviction algorithm for file cache to control disk footprint of blobfuse2
- Improved cache eviction algorithm for streamed buffers to control memory footprint of blobfuse2
- Utility to convert blobfuse CLI and config parameters to a blobfuse2 compatible config for easy migration
- CLI to mount Blobfuse2 with legacy Blobfuse config and CLI parameters (Refer to Migration guide for this)
- Version check and upgrade prompting
- Option to mount a sub-directory from a container
- CLI to mount all containers (with a allowlist and denylist) in a given storage account
- CLI to list all blobfuse2 mount points
- CLI to unmount one, multiple or all blobfuse2 mountpoints
- Option to dump logs to syslog or a file on disk
- Support for config file encryption and mounting with an encrypted config file via a passphrase (CLI or environment variable) to decrypt the config file
- CLI to check or update a parameter in the encrypted config
- Set MD5 sum of a blob while uploading
- Validate MD5 sum on download and fail file open on mismatch
- Large file writing through write streaming/Block-Cache
- 'git clone' operation is 25% faster (tested with vscode repo cloning)
- ResNet50 image classification job is 7-8% faster (tested with 1.3 million images)
- Regular file uploads are 10% faster
- Verified listing of 1-Billion files in a directory (which v1.x does not support)
You can install Blobfuse2 by cloning this repository. In the workspace root execute below commands to build the binary.
- sudo apt install fuse3 libfuse3-dev gcc
- go build -o blobfuse2
The general format of the Blobfuse2 commands is blobfuse2 [command] [arguments] --[flag-name]=[flag-value]
help
- Help about any commandmount
- Mounts an Azure container as a filesystem. The supported containers include- Azure Blob Container
- Azure Datalake Gen2 Container
mount all
- Mounts all the containers in an Azure account as a filesystem. The supported storage services includemount list
- Lists all Blobfuse2 filesystems.secure decrypt
- Decrypts a config file.secure encrypt
- Encrypts a config file.secure get
- Gets value of a config parameter from an encrypted config file.secure set
- Updates value of a config parameter.unmount
- Unmounts the Blobfuse2 filesystem.unmount all
- Unmounts all Blobfuse2 filesystems.
To see a list of commands, type blobfuse2 -h
and then press the ENTER key.
To learn about a specific command, just include the name of the command (For example: blobfuse2 mount -h
).
- Mount with blobfuse2
- blobfuse2 mount --config-file=
- Mount blobfuse2 using legacy blobfuse config and cli parameters
- blobfuse2 mountv1
- Mount all containers in your storage account
- blobfuse2 mount all --config-file=
- List all mount instances of blobfuse2
- blobfuse2 mount list
- Unmount blobfuse2
- sudo fusermount3 -u
- Unmount all blobfuse2 instances
- blobfuse2 unmount all
- Note: Blobfuse2 accepts all CLI parameters that Blobfuse does, but may ignore parameters that are no longer applicable.
- General options
--config-file=<PATH>
: The path to the config file.--log-level=<LOG_*>
: The level of logs to capture.--log-file-path=<PATH>
: The path for the log file.--foreground=true
: Mounts the system in foreground mode.--read-only=true
: Mount container in read-only mode.--default-working-dir
: The default working directory to store log files and other blobfuse2 related information.--disable-version-check=true
: Disable the blobfuse2 version check.--secure-config=true
: Config file is encrypted suing 'blobfuse2 secure` command.--passphrase=<STRING>
: Passphrase used to encrypt/decrypt config file.--wait-for-mount=<TIMEOUT IN SECONDS>
: Let parent process wait for given timeout before exit to ensure child has started.--block-cache
: To enable block-cache instead of file-cache. This works only when mounted without any config file.
- Attribute cache options
--attr-cache-timeout=<TIMEOUT IN SECONDS>
: The timeout for the attribute cache entries.--no-symlinks=true
: To improve performance disable symlink support.
- Storage options
--container-name=<CONTAINER NAME>
: The container to mount.--cancel-list-on-mount-seconds=<TIMEOUT IN SECONDS>
: Time for which list calls will be blocked after mount. ( prevent billing charges on mounting)--virtual-directory=true
: Support virtual directories without existence of a special marker blob for block blob account.--subdirectory=<path>
: Subdirectory to mount instead of entire container.--disable-compression:false
: Disable content encoding negotiation with server. If blobs have 'content-encoding' set to 'gzip' then turn on this flag.--use-adls=false
: Specify configured storage account is HNS enabled or not. This must be turned on when HNS enabled account is mounted.--cpk-enabled=true
: Allows mounting containers with cpk. Use config file or env variables to set cpk encryption key and cpk encryption key sha.
- File cache options
--file-cache-timeout=<TIMEOUT IN SECONDS>
: Timeout for which file is cached on local system.--tmp-path=<PATH>
: The path to the file cache.--cache-size-mb=<SIZE IN MB>
: Amount of disk cache that can be used by blobfuse.--high-disk-threshold=<PERCENTAGE>
: If local cache usage exceeds this, start early eviction of files from cache.--low-disk-threshold=<PERCENTAGE>
: If local cache usage comes below this threshold then stop early eviction.--sync-to-flush=false
: Sync call will force upload a file to storage container if this is set to true, otherwise it just evicts file from local cache.
- Stream options
--block-size-mb=<SIZE IN MB>
: Size of a block to be downloaded during streaming.
- Block-Cache options
--block-cache-block-size=<SIZE IN MB>
: Size of a block to be downloaded as a unit.--block-cache-pool-size=<SIZE IN MB>
: Size of pool to be used for caching. This limits total memory used by block-cache.--block-cache-path=<PATH>
: Path where downloaded blocks will be persisted. Not providing this parameter will disable the disk caching.--block-cache-disk-size=<SIZE IN MB>
: Disk space to be used for caching.--block-cache-disk-timeout=<seconds>
: Timeout for which disk cache is valid.--block-cache-prefetch=<Number of blocks>
: Number of blocks to prefetch at max when sequential reads are in progress.--block-cache-parallelism=<count>
: Number of parallel threads doing upload/download operation.--block-cache-prefetch-on-open=true
: Start prefetching on open system call instead of waiting for first read. Enhances perf if file is read sequentially from offset 0.
- Fuse options
--attr-timeout=<TIMEOUT IN SECONDS>
: Time the kernel can cache inode attributes.--entry-timeout=<TIMEOUT IN SECONDS>
: Time the kernel can cache directory listing.--negative-timeout=<TIMEOUT IN SECONDS>
: Time the kernel can cache non-existance of file or directory.--allow-other
: Allow other users to have access this mount point.--disable-writeback-cache=true
: Disallow libfuse to buffer write requests if you must strictly open files in O_WRONLY or O_APPEND mode.--ignore-open-flags=true
: Ignore the append and write only flag since O_APPEND and O_WRONLY is not supported with writeback caching.
- General options
AZURE_STORAGE_ACCOUNT
: Specifies the storage account to be connected.AZURE_STORAGE_ACCOUNT_TYPE
: Specifies the account type 'block' or 'adls'AZURE_STORAGE_ACCOUNT_CONTAINER
: Specifies the name of the container to be mountedAZURE_STORAGE_BLOB_ENDPOINT
: Specifies the blob endpoint to use. Defaults to *.blob.core.windows.net, but is useful for targeting storage emulators.AZURE_STORAGE_AUTH_TYPE
: Overrides the currently specified auth type. Case insensitive. Options: Key, SAS, MSI, SPN
- Account key auth:
AZURE_STORAGE_ACCESS_KEY
: Specifies the storage account key to use for authentication.
- SAS token auth:
AZURE_STORAGE_SAS_TOKEN
: Specifies the SAS token to use for authentication.
- Managed Identity auth:
AZURE_STORAGE_IDENTITY_CLIENT_ID
: Only one of these three parameters are needed if multiple identities are present on the system.AZURE_STORAGE_IDENTITY_OBJECT_ID
: Only one of these three parameters are needed if multiple identities are present on the system.AZURE_STORAGE_IDENTITY_RESOURCE_ID
: Only one of these three parameters are needed if multiple identities are present on the system.MSI_ENDPOINT
: Specifies a custom managed identity endpoint, as IMDS may not be available under some scenarios. Uses theMSI_SECRET
parameter as theSecret
header.MSI_SECRET
: Specifies a custom secret for an alternate managed identity endpoint.
- Service Principal Name auth:
AZURE_STORAGE_SPN_CLIENT_ID
: Specifies the client ID for your application registrationAZURE_STORAGE_SPN_TENANT_ID
: Specifies the tenant ID for your application registrationAZURE_STORAGE_AAD_ENDPOINT
: Specifies a custom AAD endpoint to authenticate againstAZURE_STORAGE_SPN_CLIENT_SECRET
: Specifies the client secret for your application registration.AZURE_STORAGE_AUTH_RESOURCE
: Scope to be used while requesting for token.
- Proxy Server:
http_proxy
: The proxy server address. Example:10.1.22.4:8080
.https_proxy
: The proxy server address when https is turned off forcing http. Example:10.1.22.4:8080
.
- CPK options:
AZURE_STORAGE_CPK_ENCRYPTION_KEY
: Customer provided base64-encoded AES-256 encryption key value.AZURE_STORAGE_CPK_ENCRYPTION_KEY_SHA256
: Base64-encoded SHA256 of the cpk encryption key.
Below diagrams guide you to choose right configuration for your workloads.
- Choose right Auth mode
- Choose right caching for Read-Only workloads
- Choose right caching for Read-Write workloads
- Choose right block-cache configuration
- Choose right file-cache configuration
- Sample File Cache Config
- Sample Block-Cache Config
- Sample Stream Config
- All Config options
- How do I generate a SAS with permissions for rename? az cli has a command to generate a sas token. Open a command prompt and make sure you are logged in to az cli. Run the following command and the sas token will be displayed in the command prompt. az storage container generate-sas --account-name --account-key -n --permissions dlrwac --start <today's date ex: 2021-03-26> --expiry
- Why do I get EINVAL on opening a file with WRONLY or APPEND flags? To improve performance, Blobfuse2 by default enables writeback caching, which can produce unexpected behavior for files opened with WRONLY or APPEND flags, so Blobfuse2 returns EINVAL on open of a file with those flags. Either use disable-writeback-caching to turn off writeback caching (can potentially result in degraded performance) or ignore-open-flags (replace WRONLY with RDWR and ignore APPEND) based on your workload.
- How to mount blobfuse2 inside a container? Refer to 'docker' folder in this repo. It contains a sample 'Dockerfile'. If you wish to create your own container image, try 'buildandruncontainer.sh' script, it will create a container image and launch the container using current environment variables holding your storage account credentials.
- Why am I not able to see the updated contents of file(s), which were updated through means other than Blobfuse2 mount?
If your use-case involves updating/uploading file(s) through other means and you wish to see the updated contents on Blobfuse2 mount then you need to disable kernel page-cache.
-o direct_io
CLI parameter is the option you need to use while mounting. Along with this, setfile-cache-timeout=0
and all other libfuse caching parameters should also be set to 0. User shall be aware that disabling kernel cache can result into more calls to Azure Storage which will have cost and performance implications.
- mkfifo : fifo creation is not supported by blobfuse2 and this will result in "function not implemented" error
- chown : Change of ownership is not supported by Azure Storage hence Blobfuse2 does not support this.
- Creation of device files or pipes is not supported by Blobfuse2.
- Blobfuse2 does not support extended-attributes (x-attrs) operations
- Blobfuse2 does not support lseek() operation on directory handles. No error is thrown but it will not work as expected.
-
Blobfuse2 does not support overlapping mount paths. While running multiple instances of Blobfuse2 make sure each instance has a unique and non-overlapping mount point.
-
Blobfuse2 does not support co-existance with NFS on same mount path. Behaviour in this case is undefined.
-
For block blob accounts, where data is uploaded through other means, Blobfuse2 expects special directory marker files to exist in container. In absence of this few file operations might not work. For e.g. if you have a blob 'A/B/c.txt' then special marker files shall exists for 'A' and 'A/B', otherwise opening of 'A/B/c.txt' will fail. Once a 'ls' operation is done on these directories 'A' and 'A/B' you will be able to open 'A/B/c.txt' as well. Possible workaround to resolve this from your container is to either
create the directory marker files manually through portal or run 'mkdir' command for 'A' and 'A/B' from blobfuse. Refer me for details on this.
-
In case of BlockBlob accounts, ACLs are not supported by Azure Storage so Blobfuse2 will by default return success for 'chmod' operation. However it will work fine for Gen2 (DataLake) accounts.
-
When Blobfuse2 is mounted on a container, SYS_ADMIN privileges are required for it to interact with the fuse driver. If container is created without the privilege, mount will fail. Sample command to spawn a docker container is
docker run -it --rm --cap-add=SYS_ADMIN --device=/dev/fuse --security-opt apparmor:unconfined <environment variables> <docker image>
By default, Blobfuse2 will log to syslog. The default settings will, in some cases, log relevant file paths to syslog. If this is sensitive information, turn off logging or set log-level to LOG_ERR.
This project is licensed under MIT.
This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.microsoft.com.
When you submit a pull request, a CLA-bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., label, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.
This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact [email protected] with any additional questions or comments.