IPFS, or the InterPlanetary File System, is a peer-to-peer decentralized protocol designed to make the web faster, safer, and more resilient by replacing or complementing traditional server-based data distribution systems like HTTP. Developed by Protocol Labs, IPFS enables files to be stored and shared directly across a distributed network of nodes rather than relying on centralized servers, enhancing data persistence and reducing vulnerability to censorship and server outages [1]. At its core, IPFS uses content addressing, where each file is assigned a unique Content Identifier (CID) based on a cryptographic hash of its contents, ensuring data integrity and immutability [2]. This contrasts with location-based addressing in HTTP, where broken links occur when servers go offline. IPFS leverages a Distributed Hash Table (DHT) to locate content across the network, allowing users to retrieve data from multiple sources simultaneously, improving speed and reliability. The protocol supports a wide range of applications, including decentralized websites, NFT metadata storage, and integration with blockchain technologies like Ethereum and Filecoin, which together form the backbone of the emerging Web3 ecosystem [3]. Despite its advantages, IPFS faces challenges related to data persistence, scalability, and privacy, which are addressed through mechanisms like pinning, remote pinning services, and integration with incentivized storage networks. Its architecture draws inspiration from systems like BitTorrent and Git, particularly through its use of a Merkle DAG for efficient data structuring, versioning, and deduplication [4]. While IPFS enhances resistance to censorship and supports open access to information, it also raises legal and ethical questions under regulations like the Digital Services Act (DSA) and the General Data Protection Regulation (GDPR)>, especially regarding the removal of illegal content and the right to be forgotten [5].
Architecture and Core Principles
The architecture and core principles of IPFS (InterPlanetary File System) are designed to create a more resilient, secure, and decentralized web by fundamentally rethinking how data is stored, addressed, and retrieved. Unlike traditional systems that rely on centralized servers and location-based addressing, IPFS implements a peer-to-peer (P2P) model where data is distributed across a global network of nodes. This design prioritizes data integrity, immutability, and resistance to censorship, forming the foundation for a new generation of web applications in the Web3 ecosystem. The system draws inspiration from established technologies like Git and BitTorrent, integrating their strengths into a unified protocol for content distribution.
Content Addressing and Cryptographic Integrity
At the heart of IPFS's architecture is the principle of content addressing, a paradigm shift from the location-based addressing used by HTTP. Instead of identifying resources by their server location (e.g., https://example.com/file.pdf), IPFS identifies them by the cryptographic hash of their content, known as a Content Identifier (CID). When a file is added to IPFS, it is processed through a cryptographic hash function, typically SHA-256, which generates a unique, fixed-length string. This hash becomes the file's CID, ensuring that any alteration to the content, no matter how minor, results in a completely different CID. This mechanism guarantees data integrity and immutability, as the CID serves as a built-in checksum that can be verified by any node in the network [6]. The CID is self-descriptive, containing metadata about the content's codec, hashing algorithm, and version (e.g., CIDv0 or CIDv1), making it portable and future-proof [7].
Merkle DAG: The Foundational Data Structure
The underlying data structure that enables IPFS's advanced capabilities is the Merkle Directed Acyclic Graph (Merkle DAG). In this model, every file and directory is broken down into smaller blocks, each of which becomes a node in the graph. Each node is identified by its own CID, which is derived from its data and the CIDs of its child nodes. This creates a hierarchical structure where the CID of a parent node depends on the CIDs of all its descendants, forming a cryptographic chain of trust. The Merkle DAG enables several critical functions: it allows for efficient data verification, as any change to a single block propagates up to the root CID; it supports deduplication, as identical blocks are stored only once regardless of how many files reference them; and it facilitates versioning, as each modification to a file creates a new root CID while preserving access to previous versions [4]. This structure is fundamental to IPFS's ability to manage large datasets, directories, and complex file relationships in a secure and efficient manner.
Distributed Hash Table and Peer Discovery
To locate and retrieve content in a decentralized network, IPFS employs a Distributed Hash Table (DHT) based on the Kademlia protocol. The DHT acts as a decentralized index that maps CIDs to the network addresses of the nodes that store them. When a user requests a file by its CID, their node queries the DHT to find peers that are providing that content. The Kademlia algorithm ensures that these queries are resolved efficiently, with the number of network hops required to find a node growing logarithmically with the size of the network. This system eliminates the need for centralized trackers or servers. Peer discovery is further enhanced through multiple mechanisms: bootstrap nodes provide initial connection points to the network; Multicast DNS (mDNS) allows for automatic discovery of peers on the same local network; and random walk strategies help new nodes populate their routing tables when starting up. This multi-layered approach ensures robust and resilient network connectivity even in highly dynamic environments [9].
Peer-to-Peer Networking and Data Transfer
IPFS operates on a pure peer-to-peer model, where every node can act as both a client and a server. This is facilitated by the libp2p networking stack, a modular framework that handles all aspects of P2P communication, including transport, security, and peer routing. When a file is shared, it is not uploaded to a central server but distributed across the network as other nodes request and cache its blocks. The Bitswap protocol governs the exchange of these blocks between peers. It operates on a "want-have" model, where nodes advertise what blocks they need and what blocks they can provide, enabling efficient, multi-source downloads. This allows users to retrieve data from the nearest or fastest available peers, improving speed and reducing latency. For a file to remain accessible, at least one node must "pin" it, preventing its deletion during garbage collection. This model shifts the responsibility for data persistence from a central authority to the network's participants [10].
Scalability, Resilience, and Trade-offs
The architecture of IPFS is designed for high resilience and horizontal scalability. The decentralized nature of the network means there are no single points of failure; if one node goes offline, the data it hosted can still be retrieved from other nodes that have it pinned. This makes IPFS highly resistant to outages and censorship, as demonstrated by instances where the network remained functional even when a significant portion of its nodes became unresponsive [11]. However, this design comes with trade-offs. The reliance on voluntary pinning means that data persistence is not guaranteed, leading to challenges with the availability of less popular content. The network also faces issues of emerging centralization, where a small percentage of nodes, often hosted on cloud infrastructure, store the majority of the data, potentially undermining the ideal of a fully distributed network [12]. To address scalability, IPFS has introduced optimizations like delegated routing, which allows lightweight nodes (e.g., in a web browser) to offload DHT queries to more powerful servers, and provide sweep, which reduces the overhead of announcing content to the DHT [13]. These innovations aim to balance the benefits of decentralization with the practical demands of a global-scale network.
Content Addressing and CID System
The Content Addressing and CID System is the foundational mechanism that distinguishes IPFS from traditional location-based web protocols like HTTP. Instead of relying on server addresses to locate data, IPFS uses a content-based identification model where each file or data block is assigned a unique, cryptographic identifier derived directly from its contents. This approach ensures data integrity, enables immutability, and supports a decentralized, resilient network architecture.
How Content Addressing Works in IPFS
In IPFS, every piece of data—whether a file, directory, or data block—is processed through a cryptographic hash function, typically SHA-256, to generate a unique fingerprint of the content [14]. This hash becomes the core of the Content Identifier (CID), a self-describing identifier that encodes not only the hash of the data but also metadata such as the hashing algorithm used, the content type, and the encoding format [7].
Unlike URLs in HTTP, which point to a specific server location (e.g., https://example.com/file.txt), a CID points to what the data is, not where it is stored. This means that identical files—regardless of their name or origin—will produce the same CID, enabling automatic deduplication across the network [16]. The content addressing model ensures that any alteration to the data, even a single bit, results in a completely different CID, making tampering immediately detectable [17].
Structure and Evolution of the CID
The CID has evolved through two primary versions: CIDv0 and CIDv1. CIDv0, the original format, uses Base58 encoding and typically begins with the prefix Qm. While simple and widely supported, CIDv0 is limited in flexibility, supporting only the SHA-256 hashing algorithm and Base58 encoding [18].
CIDv1, introduced to enhance compatibility and extensibility, supports multiple hashing algorithms (e.g., BLAKE2) and encodings (e.g., Base32), making it more suitable for integration with modern web standards and decentralized systems [19]. The self-descriptive nature of CIDv1 allows future-proofing of the identifier system, enabling support for new codecs and cryptographic primitives without breaking existing implementations [7]. Despite the advantages of CIDv1, CIDv0 remains in use for backward compatibility, and both formats coexist in the IPFS ecosystem.
Role of the Merkle DAG in Content Addressing
At the heart of IPFS’s content addressing system is the Merkle Directed Acyclic Graph (Merkle DAG), a data structure that organizes files into a hierarchical tree of blocks, each identified by its own CID [4]. In this structure, parent nodes contain references (CIDs) to their child nodes, and the hash of a parent node depends on the hashes of its children. This creates a cryptographic chain of trust, where any change in a leaf node propagates up the tree, altering the root CID [22].
The Merkle DAG enables several key functionalities:
- Efficient data verification: Users can verify the integrity of large files by checking only the root CID.
- Partial file retrieval: Clients can download only the blocks they need, improving performance.
- Versioning: Each change to a file generates a new root CID, preserving previous versions immutably.
- Shared data blocks: Multiple files can reference the same blocks, reducing redundancy and saving storage [23].
This structure is inspired by version control systems like Git, which use similar principles for tracking changes and ensuring data consistency [24].
Data Retrieval via Distributed Hash Table (DHT)
To locate content in a decentralized network, IPFS uses a Distributed Hash Table (DHT) based on the Kademlia protocol [9]. When a user requests a file by its CID, the DHT is queried to find which nodes in the network are currently storing and providing that content. The DHT maps CIDs to the Peer IDs of nodes that host the data, enabling efficient peer discovery without centralized coordination [26].
The DHT supports two types of queries:
- Content Routing: Finding nodes that provide a specific CID.
- Peer Routing: Locating the network address of a specific node by its Peer ID.
To improve performance, especially for nodes with limited resources (e.g., browsers or mobile devices), IPFS supports Delegated Routing, an HTTP-based API that allows lightweight clients to offload DHT operations to external routing servers [13]. This enhances scalability and usability without compromising the decentralized nature of the network [28].
Advantages of Content Addressing Over Location-Based Systems
Compared to traditional location-based addressing, IPFS’s content addressing offers several critical advantages:
- Immutability: Data cannot be altered without changing its CID, ensuring long-term integrity [29].
- Censorship resistance: Since data can be hosted by multiple nodes, it is difficult to remove or block [3].
- Efficiency: Identical content is stored once, reducing bandwidth and storage usage [16].
- Permanence: Links do not break even if the original host goes offline, as long as at least one node retains the data [32].
However, content addressing also introduces challenges, such as variable latency due to DHT lookups and the need for active pinning to ensure data persistence [33]. Despite these limitations, the CID system remains a cornerstone of the emerging Web3 ecosystem, enabling secure, verifiable, and decentralized data management across applications like NFT storage, decentralized websites, and blockchain-integrated systems [3].
Data Persistence and Pinning Mechanisms
Data persistence in the InterPlanetary File System (IPFS) is not guaranteed by default due to its decentralized, peer-to-peer architecture. Unlike traditional centralized storage systems where data remains available as long as the hosting server is operational, IPFS relies on a distributed network of nodes to store and serve content. The availability of any given file depends on whether at least one node in the network actively retains it. This creates a fundamental challenge: ensuring that important data remains accessible over time despite the dynamic and often transient nature of individual nodes. To address this, IPFS employs a core mechanism known as pinning, which serves as the primary method for guaranteeing data persistence. Without explicit pinning, files are subject to automatic removal through a process called garbage collection, which clears temporary data to free up storage space on a node. Therefore, the longevity of content on IPFS is directly tied to the pinning practices of its users and the broader ecosystem of storage providers [35].
The Role and Function of Pinning
Pinning is the process by which a user or a node explicitly instructs the IPFS system to retain a specific file or directory, identified by its Content Identifier (CID), and prevent it from being deleted during garbage collection. When a file is first added to an IPFS node, it is stored in a temporary cache. The node does not automatically assume responsibility for long-term storage. By issuing a ipfs pin add <CID> command, the user marks the content as "pinned," signaling to the node that this data is essential and must be preserved [36]. This mechanism shifts the responsibility of data persistence from the network as a whole to individual participants. The fundamental principle is that a file will remain accessible on the network as long as at least one node somewhere in the world continues to pin it. This creates a model of "voluntary persistence," where the durability of data is a function of community interest and active maintenance. For example, highly popular content, such as the code for a widely used open-source project or the metadata for a popular NFT, is likely to be pinned by many nodes, ensuring its robust availability. Conversely, less popular or niche content is at a higher risk of becoming inaccessible if its original publisher or a few dedicated supporters cease to pin it, leading to what is known as "link rot" in the decentralized web [37].
Remote Pinning Services and Managed Solutions
To overcome the limitations of relying on personal nodes, which may be offline or have limited storage, the IPFS ecosystem has developed a robust market for remote pinning services. These are third-party providers that offer reliable, always-on infrastructure for pinning content, ensuring high availability and durability. These services are essential for production-grade applications, such as hosting the frontend of a dApp or storing critical NFT assets. Key providers include Pinata, Filebase, Aleph Cloud, and Infura, each offering APIs and user-friendly dashboards for managing pinned content [38]. These services often provide additional features such as dedicated IPFS Gateway access, content monitoring, and analytics. The use of these managed solutions has become a standard practice in the Web3 development workflow, as they abstract away the complexity of node management. They operate using the standardized IPFS Pinning Service API, which ensures interoperability and allows developers to switch providers or use multiple services for redundancy without changing their application code [39]. This ecosystem of services transforms IPFS from a purely peer-to-peer network into a more practical and reliable storage platform, bridging the gap between the ideal of decentralization and the real-world need for guaranteed uptime.
Integration with Incentivized Storage: The Role of Filecoin
While remote pinning services solve the availability problem, they often rely on centralized providers, which can reintroduce points of failure and trust. To create a truly decentralized and economically sustainable model for persistent storage, IPFS is integrated with Filecoin, a blockchain-based protocol that creates a marketplace for storage. Filecoin addresses the core limitation of IPFS by introducing financial incentives for data persistence. In this model, users who need long-term storage make "storage deals" with storage providers (miners) by paying in the native FIL token. The Filecoin network then uses cryptographic proofs, specifically Proof of Replication (PoRep) and Proof of Spacetime (PoSt), to verify that the miners are actually storing the data correctly and continuously over time [40]. This integration means that data stored on IPFS can be guaranteed to persist for a specified duration, with verifiable proof. Services like Web3.Storage and Filecoin Pin simplify this process by allowing users to store data with a single API call, automatically handling both the IPFS upload and the Filecoin storage deal [41]. This synergy creates a powerful two-layer architecture: IPFS provides an efficient, content-addressed network for data distribution, while Filecoin provides a verifiable, incentivized layer for permanent storage, forming a comprehensive solution for decentralized data management.
Challenges and Best Practices for Ensuring Data Durability
Despite the availability of pinning and incentivized storage, ensuring data durability on IPFS presents ongoing challenges. A significant issue is the observed trend toward centralization within the network, where a small number of nodes, often operated by large cloud providers, host the majority of the content. This undermines the resilience of the network, as it creates potential single points of failure [12]. Another challenge is the low rate of natural data replication; studies show that only a small fraction of files are replicated more than a few times, making many datasets vulnerable [43]. To mitigate these risks, best practices for developers and organizations include using multiple pinning services for redundancy, integrating with Filecoin for critical data, and monitoring the health and availability of their CIDs. Tools like IPFS Cluster allow for the orchestration of pinning across a group of nodes, providing high availability and automated management for large-scale deployments [44]. Ultimately, the responsibility for data persistence in IPFS is a shared one, requiring a combination of technical tools, economic incentives, and community participation to build a truly resilient and permanent web.
Integration with Blockchain and Web3
The integration of the InterPlanetary File System (IPFS) with blockchain technologies and the broader Web3 ecosystem represents a foundational shift in how data is stored, accessed, and verified in decentralized applications. By combining IPFS’s efficient, content-based file distribution with the immutability and trustless verification of blockchain, developers can build resilient, censorship-resistant, and scalable applications that redefine digital ownership and data integrity. This synergy addresses key limitations of traditional web architectures, particularly the high cost and inefficiency of storing large data directly on-chain, by enabling off-chain storage with on-chain verification.
Data Storage and Off-Chain Data Management
One of the most significant contributions of IPFS to blockchain systems is its role in managing large-scale data off-chain. Blockchains such as Ethereum are designed for secure, tamper-proof transaction recording but are not optimized for storing large files like images, videos, or documents due to high gas fees and scalability constraints. IPFS solves this by allowing such data to be stored off-chain while only the unique Content Identifier (CID) is recorded on the blockchain. This approach ensures that the data remains verifiable and immutable, as any alteration to the file would change its CID, breaking the cryptographic link [45].
This model is widely adopted in NFT ecosystems, where digital art, metadata, and associated media are stored on IPFS. Platforms like NFT.Storage, Pinata, and Filebase provide dedicated services for uploading and pinning NFT assets, ensuring long-term availability and integrity. The CID is then embedded into the NFT’s smart contract, creating a permanent, decentralized reference to the underlying content [46]. This prevents issues like "link rot" that plague centralized hosting solutions, where the removal or relocation of a file breaks the connection between the token and its representation.
Smart Contracts and Decentralized Application (dApp) Frontends
IPFS is also instrumental in hosting the frontends of decentralized applications (dApps). Traditional web applications rely on centralized servers, making them vulnerable to downtime, censorship, and single points of failure. In contrast, dApp user interfaces hosted on IPFS are distributed across a global network of nodes, ensuring high availability and resistance to censorship. For example, many Ethereum-based dApps use IPFS to serve their web interfaces, allowing users to interact with the application even if the original developers go offline [47].
This integration enhances the trustless nature of Web3 by ensuring that both the backend logic (via smart contracts) and the frontend interface (via IPFS) are decentralized. Users can verify that the UI they are interacting with has not been tampered with by checking its CID, which can be published through trusted channels or embedded in governance systems. This level of transparency and verifiability is critical for maintaining user trust in decentralized finance (DeFi), governance platforms, and digital identity systems.
Interoperability with Filecoin for Persistent Storage
While IPFS excels at content distribution, it does not inherently guarantee data persistence, as files are only available as long as at least one node is actively pinning them. This limitation is addressed through integration with Filecoin, a decentralized storage network built on top of IPFS. Filecoin introduces a market-based incentive model where users pay in FIL tokens to have their data stored reliably by miners who must prove ongoing data custody through cryptographic proofs such as Proof of Replication (PoRep) and Proof of Spacetime (PoSt) [40].
This combination creates a robust ecosystem: IPFS handles fast, efficient content addressing and retrieval, while Filecoin ensures long-term, verifiable storage. Tools like Filecoin Pin automate the process of backing up IPFS content on the Filecoin network, enabling developers to build applications with guaranteed data durability. Services such as web3.storage and Lighthouse Storage further simplify this integration by offering user-friendly APIs that manage both IPFS pinning and Filecoin storage deals, making persistent decentralized storage accessible even to non-technical users [49].
Governance, Identity, and Decentralized Systems
Beyond storage, IPFS plays a crucial role in decentralized governance and identity systems. Platforms like Snapshot, used for voting in decentralized autonomous organizations (DAOs), store proposal data and voting records on IPFS to ensure transparency and immutability. This allows participants to independently verify the integrity of governance processes without relying on centralized servers [50].
Similarly, IPFS supports the development of self-sovereign identity (SSI) frameworks by enabling secure, decentralized storage of identity credentials and verifiable claims. When combined with blockchain-based identity solutions, IPFS allows individuals to maintain control over their personal data, sharing only what is necessary with verifiable authenticity. This aligns with the core principles of Web3: user empowerment, data ownership, and resistance to surveillance and censorship.
Challenges and Emerging Solutions
Despite its advantages, the integration of IPFS with blockchain and Web3 faces challenges related to data persistence, performance, and governance. The reliance on voluntary pinning means that without economic incentives or managed services, critical data may become inaccessible. Moreover, the lack of native privacy in IPFS—where all content is public by default—requires additional layers of end-to-end encryption to protect sensitive information [51].
Efforts to address these issues include the development of federated or private IPFS networks for enterprise use, automated replication policies using machine learning, and hybrid models that combine decentralized storage with traditional CDN-like performance optimizations. Additionally, emerging governance frameworks aim to establish clear responsibilities and moderation mechanisms for handling illegal content, balancing freedom of expression with legal compliance in line with regulations like the Digital Services Act (DSA) [52].
In conclusion, IPFS serves as a critical infrastructure layer for the Web3 vision, enabling scalable, secure, and decentralized data management across blockchain applications. Its integration with technologies like Ethereum and Filecoin not only enhances the functionality of dApps and NFTs but also advances the broader goals of digital sovereignty and open access. As the ecosystem matures, continued innovation in persistence, privacy, and governance will be essential to realizing the full potential of a truly decentralized web.
Network Resilience and Decentralization
The InterPlanetary File System (IPFS) is fundamentally designed to enhance network resilience and decentralization by replacing the traditional client-server model of the web with a peer-to-peer (P2P) architecture. Unlike centralized systems such as HTTP, where data is hosted on specific servers and becomes inaccessible if those servers go offline, IPFS distributes content across a global network of nodes. This architectural shift eliminates single points of failure, making the network inherently more robust and resistant to outages, censorship, and distributed denial-of-service (DDoS) attacks [1].
In IPFS, every file is identified by a unique Content Identifier (CID) derived from its cryptographic hash. This content addressing mechanism ensures that data can be retrieved from any node that holds a copy, rather than relying on a fixed location. As long as at least one node in the network stores and shares the content, it remains accessible. This model significantly increases data persistence and availability, even in the face of partial network failures or targeted takedowns. For example, when Wikipedia was blocked by the Turkish government, it was successfully republished on IPFS, ensuring continued access to information through decentralized hosting [54].
Peer Discovery and Network Topology
IPFS employs a sophisticated set of mechanisms to enable nodes to discover and connect with each other, ensuring the network remains functional and adaptive even under high node churn. Upon initialization, a node connects to a predefined list of bootstrap nodes, which serve as entry points into the network and help the new node discover other peers. This process is facilitated by the libp2p framework, a modular networking stack that underpins IPFS and supports various discovery protocols [55].
In local networks, IPFS uses Multicast DNS (mDNS) to allow nodes to automatically detect each other without manual configuration. For global discovery, IPFS relies on a Distributed Hash Table (DHT) based on the Kademlia protocol. The DHT maps CIDs to the network addresses of nodes that store them, enabling efficient content routing without centralized directories. To accelerate discovery when a node’s routing table is empty, IPFS implements a random walk strategy, where the node performs random queries on the DHT to populate its peer list quickly [56].
This multi-layered discovery system ensures that IPFS can maintain connectivity and content availability even when large portions of the network become unresponsive. A notable incident in 2023 demonstrated this resilience: despite 60% of DHT nodes becoming unreachable, the majority of content remained accessible, with only a modest increase in retrieval latency [11].
Decentralized Routing and Scalability
The routing architecture of IPFS distinguishes it from other P2P systems like BitTorrent and Freenet. While BitTorrent depends on centralized trackers or a less structured DHT for peer coordination, IPFS uses a fully decentralized, structured DHT optimized for content discovery. This allows IPFS to support universal content addressing via CIDs, enabling features such as global deduplication—where identical files are stored only once—and persistent links that do not break when content moves [58].
Compared to Freenet, which prioritizes anonymity and censorship resistance through encrypted, distributed storage, IPFS emphasizes efficiency and scalability in content distribution. It does not provide built-in privacy, but its use of Kademlia-based DHT enables faster and more targeted content lookups, making it better suited for integration with web applications and blockchain technologies [59].
To address scalability challenges, IPFS has introduced several optimizations. The Provide Sweep mechanism, introduced in Kubo v0.39, reduces the number of DHT lookups required to announce content availability by up to 97%, greatly improving efficiency for nodes hosting large datasets [60]. Additionally, delegated routing allows lightweight nodes, such as those running in web browsers, to offload routing operations to external servers via HTTP APIs. This enhances performance without compromising decentralization, as nodes retain control over their data [13].
Resilience in Dynamic and Adversarial Environments
IPFS’s network resilience is particularly evident in highly dynamic environments where nodes frequently join and leave the network. The DHT’s ability to adapt incrementally to changes in topology ensures that the system remains functional even with high node turnover. Studies estimate that around 44,474 nodes are active on the IPFS network at any given time, many operating behind NATs, contributing to a highly distributed and censorship-resistant infrastructure [62].
However, challenges remain. Despite its decentralized design, IPFS exhibits signs of emergent centralization, with approximately 5% of nodes hosting over 80% of the content, often due to reliance on cloud-based infrastructure [63]. This concentration can undermine resilience by creating de facto dependencies on a few powerful nodes. Nevertheless, the distributed nature of the DHT and the redundancy provided by content replication help mitigate the risks of Sybil attacks and single points of failure.
Furthermore, IPFS integrates with complementary technologies to enhance robustness. For instance, IPFS Cluster enables coordinated pinning across multiple nodes, ensuring high availability and automated redundancy. Similarly, integration with Filecoin introduces economic incentives for long-term data storage, addressing the issue of content persistence through verifiable proofs of replication and spacetime [40].
In summary, IPFS achieves a high degree of network resilience and decentralization through a combination of content addressing, peer-to-peer networking, and advanced routing protocols. While challenges such as replication imbalance and emergent centralization persist, ongoing improvements in routing efficiency, delegated operations, and incentive layers continue to strengthen its capacity to support a durable, censorship-resistant web.
Scalability Challenges and Performance Optimization
The InterPlanetary File System (IPFS) offers a transformative model for decentralized data distribution, but its scalability and performance face significant challenges when deployed at large scale. While the protocol excels in resilience and censorship resistance, its underlying architecture introduces bottlenecks that affect efficiency, availability, and user experience. These challenges stem from structural limitations in content discovery, data replication, and network topology, but a range of optimization strategies—both technical and architectural—are being developed and deployed to address them.
Centralization and Uneven Data Distribution
Despite its decentralized design, IPFS exhibits a strong tendency toward centralization in practice. Empirical studies reveal that over 80% of content on the network is hosted by less than 5% of nodes, many of which are operated by cloud providers [12]. This concentration undermines the core principle of decentralization, creating de facto single points of failure and reducing the geographic distribution of content. The reliance on centralized cloud infrastructure for node operation introduces risks similar to those in traditional web architectures, including potential censorship and service disruption.
This uneven distribution is exacerbated by the lack of built-in incentives for nodes to host content. Unlike systems such as Filecoin, which use economic rewards to encourage storage, IPFS relies on voluntary pinning. As a result, only popular or strategically important content tends to be widely replicated, while less popular data becomes increasingly difficult to retrieve.
Limited Data Replication and Content Availability
A critical scalability issue in IPFS is the low rate of natural data replication. Research indicates that only 2.71% of files are replicated more than five times across the network [43]. This scarcity of redundancy poses a serious threat to data durability, especially for content that is not frequently accessed. If the few nodes hosting a particular file go offline, the content becomes effectively lost, even though its Content Identifier (CID) remains valid.
This problem is compounded by the absence of guaranteed persistence. Unlike traditional storage systems that offer service-level agreements, IPFS provides no assurance that data will remain available over time. The reliance on manual or service-based pinning means that data longevity depends on the continued operation of specific nodes or third-party providers, such as Pinata or Filebase, rather than on the network as a whole.
DHT Bottlenecks and Routing Overhead
The Distributed Hash Table (DHT), which IPFS uses to locate content and peers, is a major source of performance bottlenecks. Based on the Kademlia protocol, the DHT enables decentralized content discovery by mapping CIDs to the nodes that store them [9]. However, the process of "providing" a CID—announcing to the network that a node hosts a particular piece of content—requires numerous DHT lookups, creating significant computational and network overhead.
This overhead becomes especially problematic for nodes that self-host large volumes of data, such as those used in enterprise or archival applications. The inefficiency of the DHT can lead to slow content discovery, increased latency, and higher resource consumption, limiting the protocol's ability to scale to millions of nodes and petabytes of data.
Performance Limitations in Resource-Constrained Environments
IPFS performance is particularly constrained in resource-limited environments, such as web browsers or mobile devices. Running a full IPFS node in these contexts is often impractical due to memory, bandwidth, and processing power limitations. The complexity of maintaining DHT routing tables and managing peer connections makes it difficult to achieve acceptable performance without offloading these tasks.
To address this, the concept of delegated routing has been introduced. This allows lightweight clients to delegate DHT operations to external servers via standardized HTTP APIs [13]. By offloading routing to more powerful nodes, browsers and mobile apps can interact with IPFS efficiently, improving user experience and enabling broader adoption of decentralized applications (dApps) [28].
Optimization Strategies and Emerging Solutions
Several innovative strategies have been developed to overcome IPFS's scalability and performance limitations:
-
Provide Sweep: Introduced in Kubo v0.39, this optimization reduces the number of DHT lookups required to announce content by grouping CIDs assigned to the same DHT servers. This can decrease lookup operations by up to 97%, dramatically improving the efficiency of large-scale data hosting [60].
-
IPFS Cluster and Elastic IPFS: These tools provide coordinated management of multiple IPFS nodes, enabling automated replication, load balancing, and high availability. IPFS Cluster is particularly useful for organizations that need to ensure data redundancy and persistent access, while Elastic IPFS offers cloud-native scalability [71].
-
Integration with Advanced Networking: Experiments with next-generation network architectures like SCION have shown that replacing traditional TCP/IP with more secure and predictable routing can improve IPFS content retrieval times by up to 2.9 times [72].
-
Bitswap Protocol Enhancements: Ongoing improvements to the Bitswap protocol, which governs block exchange between peers, have optimized the transfer of large files and container images, reducing latency and improving bandwidth utilization [73].
-
Hybrid Models with Blockchain: To ensure long-term data persistence, IPFS is increasingly combined with blockchain-based incentive layers. For example, smart contracts on Ethereum can be used to reward nodes for pinning critical data, creating a market-driven approach to content availability [74].
These advancements demonstrate that while IPFS faces significant scalability hurdles, the ecosystem is actively evolving to meet them. Through a combination of protocol-level optimizations, architectural innovations, and integration with complementary technologies, IPFS is moving toward a more efficient, resilient, and scalable future for decentralized data storage and distribution.
Privacy, Security, and Legal Implications
The adoption of decentralized technologies like the InterPlanetary File System brings transformative potential for data resilience and censorship resistance, but it also introduces complex challenges in the realms of privacy, security, and legal compliance. Unlike traditional centralized systems, where a single entity governs data access and moderation, IPFS operates on a peer-to-peer model that distributes control across a global network of nodes. This architectural shift fundamentally alters the dynamics of content availability, user accountability, and regulatory enforcement, creating a tension between technological innovation and established legal frameworks.
Privacy Risks and User Anonymity
IPFS, in its native form, does not guarantee user privacy. All content stored on the network is publicly accessible to anyone who possesses the corresponding Content Identifier (CID), which acts as a permanent, content-based address. This transparency ensures data integrity but exposes significant privacy risks, particularly when sensitive or personal information is inadvertently shared. Once uploaded, such data can be indexed by tools like IPFS-search and remain permanently available across the network, leading to potential violations of privacy rights [75].
Furthermore, the network's structure allows for the tracking of node activities. Projects such as IPFS-CID-Hoarder can monitor which nodes are pinning specific content, enabling the creation of user profiles based on their data-sharing behavior [76]. This capability raises concerns about surveillance and the potential for re-identification, even in a decentralized environment. While IPFS nodes are identified by public PeerIDs, the combination of this identifier with behavioral data can compromise user anonymity, especially in jurisdictions with advanced monitoring capabilities [77].
To mitigate these risks, users and developers are advised to implement end-to-end encryption before uploading data to IPFS. By encrypting files with robust algorithms such as AES or Elliptic Curve Cryptography (ECC), only authorized parties with the decryption key can access the content, transforming IPFS into a secure storage layer [78]. Additionally, the use of private or federated IPFS networks can restrict participation to trusted entities, enhancing control over data exposure [51].
Security Challenges and Malicious Content
While IPFS provides strong guarantees of data integrity through its use of cryptographic hashing and the Merkle DAG structure, it does not inherently prevent the distribution of malicious content. The network has been exploited to host malware, phishing kits, and credential harvesters, often leveraging the difficulty of content removal to evade detection [80]. The immutability of content, a feature designed to resist censorship, becomes a liability when applied to illegal or harmful material.
The persistence of such content is further enabled by the "pinning problem": a file remains accessible as long as at least one node in the network continues to pin it. This makes complete eradication nearly impossible, as it would require the coordinated removal of the content from every hosting node. Studies have shown that IPFS is increasingly used to distribute child sexual abuse material (CSAM), including content generated by artificial intelligence, posing significant challenges for law enforcement and child protection agencies [81].
To address these threats, the IPFS community has introduced mechanisms like the "Bad Bits Denylist," a shared list of CIDs associated with known malicious content. Gateways such as ipfs.io and dweb.link can use this list to block access to harmful files, although this does not remove them from the network itself [82]. Research is also ongoing into automated detection systems that use machine learning to identify suspicious patterns in node behavior or content metadata, aiming to flag and isolate malicious activity in real time [83].
Legal Implications and Regulatory Compliance
The decentralized nature of IPFS places it in a legal gray area, particularly under the European Union's regulatory framework. The Digital Services Act (DSA), which governs online intermediaries, is designed for centralized platforms where a single entity can be held accountable for content moderation. IPFS, however, lacks a central provider, making it difficult to apply traditional liability models. As a result, IPFS as a protocol does not neatly fit into the DSA's categories of "mere conduit," "caching," or "hosting," and its status as a "neutral intermediary" remains legally ambiguous [5].
A critical conflict arises with the General Data Protection Regulation (GDPR), specifically its "right to be forgotten" (Article 17). The immutability and distributed replication of data in IPFS are fundamentally at odds with the GDPR's requirement for data erasure. Once personal data is published on IPFS, it cannot be reliably deleted from all nodes, creating a structural incompatibility that challenges the enforceability of EU data protection laws [85].
Efforts to resolve this tension include the development of hybrid systems that combine IPFS with blockchain-based access control. These frameworks use smart contracts to manage encryption keys and revoke access to data, providing a technical workaround for data deletion obligations [86]. However, such solutions depend on user compliance and do not eliminate the underlying data from the network, leaving legal uncertainties unresolved.
Governance and the Future of Content Moderation
The governance of IPFS is evolving to address these challenges. While the protocol was designed to resist censorship, its potential for abuse necessitates new models of decentralized moderation. Inspired by the Fediverse, proposals such as FedMod suggest distributed moderation systems where nodes collaboratively flag and filter harmful content based on reputation or consensus [87]. These approaches aim to balance freedom of expression with community safety, avoiding the pitfalls of both unchecked anarchy and centralized control.
The role of gateway providers is also crucial. Centralized gateways like those operated by Cloudflare or Pinata can implement content filtering and respond to takedown requests, acting as de facto moderators. This creates a paradox: while the underlying network is decentralized, user access often depends on centralized services that can be pressured by legal authorities [88].
In conclusion, IPFS presents a profound reconfiguration of the digital landscape, offering unprecedented resilience and freedom but also introducing significant risks to privacy, security, and legal order. Its success in the long term will depend not only on technological innovation but on the development of robust, inclusive governance models that can navigate the complex interplay between decentralization, human rights, and the rule of law.
Use Cases and Practical Applications
The InterPlanetary File System (IPFS) has evolved beyond its foundational role as a decentralized file-sharing protocol to become a critical infrastructure component in the emerging Web3 ecosystem. Its unique architecture, based on content addressing and peer-to-peer distribution, enables a wide range of practical applications that leverage its core strengths: resilience, immutability, censorship resistance, and efficient data distribution. These use cases span from decentralized websites and digital art preservation to enterprise document management and innovative Web3 platforms, demonstrating IPFS's versatility in addressing the limitations of traditional, centralized web technologies.
Hosting of Static Websites and Decentralized User Interfaces
One of the most widespread and impactful applications of IPFS is the hosting of static websites. Unlike traditional hosting that relies on centralized servers, IPFS distributes website files across a global network of nodes. This architecture eliminates single points of failure, making sites highly resistant to downtime caused by server outages or DDoS attacks. Static sites, such as personal blogs, portfolios, and technical documentation, are particularly well-suited for IPFS because they do not require dynamic server-side processing. Tools like ipfs-deploy, GitHub Actions, and dedicated services such as Pinata and Filebase have streamlined the process of publishing a site to IPFS, allowing developers to deploy with a single command. Once published, the site is accessible via a unique Content Identifier (CID) through public IPFS gateways like ipfs.io or private, dedicated gateways. To provide a user-friendly, mutable URL for a site that may be updated, developers use the InterPlanetary Naming System (IPNS), which allows a human-readable name to point to the latest CID of the website, ensuring a persistent link despite content changes [89].
This capability extends to the frontends of decentralized applications (dApps). Many dApps, especially those built on the Ethereum blockchain, use IPFS to host their user interfaces. This practice ensures that the application's UI remains accessible even if the project's backend services or traditional web servers go offline, which is crucial for maintaining trust and usability in a decentralized context. By decoupling the UI from centralized infrastructure, developers build applications that are more resilient and resistant to censorship, a fundamental principle of the Web3 vision [47].
Archival and Censorship-Resistant Data Storage
IPFS's ability to create permanent, immutable links to data makes it an ideal platform for archival and censorship-resistant storage. The most prominent example is the hosting of Wikipedia on IPFS. When access to Wikipedia was blocked by the Turkish government in 2017, the site was mirrored on IPFS, allowing users to continue accessing the information through the decentralized network. This demonstrated IPFS's power as a tool for preserving access to knowledge in the face of government censorship [54].
This use case is vital for journalists, activists, and organizations operating in repressive regimes. By publishing sensitive documents, investigative reports, or historical records on IPFS, they can ensure that the information persists and remains verifiable, even if the original source is taken down. The integrity of the data is guaranteed by its cryptographic hash (CID), meaning any attempt to alter the content would result in a completely different identifier, making tampering easily detectable. This has led to the creation of digital archives and libraries that are designed to be permanent and resistant to both censorship and the natural decay of digital media, often referred to as the "link rot" problem that plagues the traditional web [3].
NFT Data and Metadata Storage
The rise of non-fungible tokens (NFTs) has cemented IPFS as a standard for digital asset storage. NFTs are blockchain-based tokens that represent ownership of a unique digital item, such as artwork, music, or collectibles. However, storing the actual media files (images, videos, 3D models) directly on a blockchain is prohibitively expensive and inefficient. Instead, the NFT's smart contract typically stores only a link to the file's location. Using a traditional HTTP link is risky, as the image could disappear if the hosting server goes down, rendering the NFT worthless—a phenomenon known as "link rot."
IPFS solves this problem perfectly. By storing the NFT's media and its associated metadata (which describes the asset's properties) on IPFS, a permanent and verifiable link is created. The smart contract stores the file's CID, which guarantees that the content is exactly as it was when the NFT was minted. This ensures the long-term persistence and integrity of the digital asset. Services like NFT.Storage, Pinata, and Venly have emerged to provide specialized, user-friendly platforms for developers and artists to upload and pin their NFT data to IPFS, often with integrated support for the Filecoin network to guarantee long-term storage [46]. This integration of IPFS with blockchain technology is a cornerstone of the digital collectibles and art market, providing a secure and reliable foundation for digital ownership.
Enterprise and Commercial Applications
Beyond the Web3 sphere, IPFS is being adopted by enterprises for its security, efficiency, and global accessibility. In the field of international trade and logistics, companies leverage IPFS to store and share critical shipping and customs documents. For example, Morpheus.Network uses IPFS to create a secure, globally accessible record of trade documentation, reducing the risk of fraud and streamlining processes [94]. Similarly, CargoX combines IPFS with NFTs to manage digital bills of lading, enabling faster and more secure transfer of ownership rights for physical goods [95].
Italian company Verifica integrates IPFS to enhance data traceability and certification, promoting sustainable and low-energy solutions for data management [96]. These applications benefit from IPFS's ability to provide a tamper-proof, auditable history of data. The use of content addressing ensures that any document retrieved from the network is identical to the original, which is essential for legal and compliance purposes. This enterprise adoption demonstrates that the value of decentralized, content-addressed storage extends far beyond the cryptocurrency community and into core business operations.
Peer-to-Peer File Sharing and Innovative Platforms
At its core, IPFS enables efficient and secure peer-to-peer (P2P) file sharing. Users can install a local IPFS node, import folders, and share public links based on the CID. As long as at least one node on the network is "pinning" (actively storing) the content, it remains accessible to anyone with the link. This model is ideal for creating community-driven archives, distributing open-source software, or performing distributed backups.
This foundational capability has inspired a new generation of innovative platforms. Cloudest is a decentralized cloud storage platform that combines IPFS with the Ethereum blockchain to offer secure and transparent storage services [97]. Another project, IP5, aims to integrate IPFS with blockchain and artificial intelligence to manage digital identities and verified services. These projects highlight how IPFS is not just a storage solution but a building block for an entire ecosystem of decentralized services, fostering a more open, secure, and user-controlled internet [3].
Integration with Blockchain and Web3 Infrastructure
The synergy between IPFS and blockchain technology is a defining characteristic of the Web3 landscape. While blockchains like Ethereum are excellent for recording immutable transactions and smart contract logic, they are ill-suited for storing large amounts of data due to cost and scalability constraints. IPFS provides the perfect off-chain solution. By storing the bulk of the data (e.g., a dApp's UI, NFT metadata, or large datasets) on IPFS and only recording its CID on the blockchain, developers create hybrid systems that are both efficient and secure. This integration is used in diverse applications, from managing digital document repositories to powering smart contract-based platforms [99]. The combination of IPFS for data storage and blockchain for transactional integrity forms the backbone of many Web3 applications, enabling a new paradigm of trustless, decentralized computing.
Governance and Future Development
The governance and future development of the InterPlanetary File System (IPFS) are shaped by a complex interplay of technical innovation, community collaboration, and evolving regulatory challenges. As a decentralized protocol, IPFS does not operate under a traditional corporate or centralized authority. Instead, its direction is guided by Protocol Labs, the organization that originally developed the technology, in conjunction with a global community of developers, researchers, and ecosystem participants. This hybrid model of governance combines top-down technical leadership with bottom-up community input, aiming to balance rapid innovation with open, transparent decision-making [100].
Governance Model and Community Involvement
IPFS’s governance is inherently sociotechnical, relying on both technical standards and social coordination. While Protocol Labs maintains significant influence over the protocol’s roadmap and core implementations like Kubo, the project actively encourages community participation through public forums, GitHub repositories, and open RFCs (Request for Comments). This collaborative approach mirrors the governance models of other open-source and decentralized technologies such as Linux or Git, where contributions and consensus drive evolution [101]. The project has also initiated discussions on formalizing governance structures through independent foundations to ensure long-term sustainability and reduce reliance on a single entity [102].
A critical aspect of governance involves managing the tension between decentralization and practical control. Although IPFS is designed to be a peer-to-peer network without central points of failure, the reality is that key infrastructure components—such as bootstrap nodes, public gateways, and pinning services—are often operated by a small number of organizations, including Protocol Labs, Pinata, and Cloudflare. This concentration raises concerns about de facto centralization, where a few actors can influence network behavior, moderate content, or shape technical standards [12]. For example, public gateways like ipfs.io or dweb.link can implement denylists to block access to specific CIDs, effectively enabling a form of content moderation despite the protocol’s decentralized architecture [104].
Legal and Regulatory Challenges
The decentralized nature of IPFS places it in a legal gray area, particularly under frameworks like the Digital Services Act (DSA) in the European Union. The DSA imposes responsibilities on digital intermediaries for content moderation, transparency, and user protection, but it was designed with centralized platforms in mind. IPFS, as a protocol rather than a service, does not fit neatly into existing categories such as "mere conduit," "caching," or "hosting." This ambiguity raises questions about whether IPFS can be considered a neutral intermediary under the DSA’s safe harbor provisions [5]. While individual nodes may act passively, the network’s immutability and persistence of data conflict with legal requirements such as the right to be forgotten under the General Data Protection Regulation (GDPR), creating a structural tension between technological design and regulatory compliance [85].
Moreover, the inability to remove illegal content—such as child sexual abuse material, terrorist propaganda, or pirated works—poses significant ethical and legal challenges. Unlike centralized platforms that can swiftly delete or block content, IPFS lacks native mechanisms for content removal. Instead, mitigation relies on voluntary actions, such as nodes refusing to pin certain CIDs or gateways implementing blocklists. Projects like the Bad Bits Denylist provide shared databases of harmful CIDs, enabling coordinated filtering while preserving the underlying network’s openness [82]. However, these measures are inherently limited and highlight the need for new governance models that support decentralized content moderation, potentially inspired by systems in the Fediverse or using federated learning for distributed trust [87].
Future Development and Technical Roadmap
The future development of IPFS is focused on addressing scalability, performance, and usability challenges to enable broader adoption in enterprise and mainstream applications. One of the most significant technical hurdles is the inefficiency of the Distributed Hash Table (DHT) at scale, particularly the overhead associated with announcing content availability ("provide" operations). The Provide Sweep optimization, introduced in Kubo v0.39, reduces DHT lookup operations by up to 97% by batching CID announcements, significantly improving performance for nodes hosting large datasets [60]. Similarly, Delegated Routing allows lightweight clients, such as web browsers, to outsource DHT queries to external servers via HTTP APIs, enhancing accessibility without compromising decentralization [13].
Scalability is further enhanced through tools like IPFS Cluster, which orchestrates pinning and replication across multiple nodes, ensuring high availability and redundancy. For cloud-native environments, solutions like Elastic IPFS enable dynamic scaling of node clusters based on demand, supporting large-scale deployments in edge computing and content delivery networks (CDN) [111]. Integration with next-generation networking protocols such as SCION has shown potential to improve data retrieval speeds by up to 2.9 times compared to traditional TCP/IP, offering a path toward more resilient and performant data transport [72].
Another key area of development is the integration of IPFS with blockchain ecosystems, particularly through its synergy with Filecoin. While IPFS excels at content addressing and distribution, it does not natively incentivize long-term data storage. Filecoin addresses this gap by creating a decentralized marketplace where users pay in FIL tokens to have their data stored and verified over time. This combination enables verifiable, persistent storage, making the duo a foundational layer for Web3 applications [40]. Tools like Filecoin Pin automate the process of backing up IPFS content on the Filecoin network, ensuring durability while simplifying the developer experience [114].
Sociotechnical Impacts and Ethical Considerations
The widespread adoption of IPFS could have profound sociotechnical impacts, reshaping access to information, digital sovereignty, and the power dynamics of the internet. By enabling censorship-resistant publishing, IPFS empowers activists, journalists, and marginalized communities to share information without fear of takedown [3]. Projects like Akasha, a decentralized social network, demonstrate how IPFS can support user-owned platforms that resist corporate control [116].
However, this same resilience raises ethical dilemmas. The permanence of data on IPFS conflicts with societal norms around privacy, accountability, and the right to be forgotten. Once data is published, it can persist indefinitely, even if later deemed harmful or inaccurate. This challenge is compounded by the difficulty of detecting and removing illegal content, as automated systems struggle to scan decentralized networks effectively [52]. Future governance models may need to incorporate privacy-preserving technologies such as end-to-end encryption, zero-knowledge proofs, or private query protocols like Peer2PIR to allow secure and confidential access to data [118].
Furthermore, while IPFS aims to reduce digital inequalities by enabling peer-to-peer information sharing, its technical complexity and reliance on high-bandwidth infrastructure risk excluding users in low-resource regions. Efforts to bridge this gap include the development of lightweight clients, offline-first applications, and community-run mesh networks. The long-term success of IPFS will depend not only on technical innovation but also on inclusive governance, ethical foresight, and the ability to navigate an increasingly complex regulatory landscape.