Chapter XIII: DePIN
The infrastructure you depend on (cell towers, cloud servers, street maps) has always been built the same way: a corporation deploys hardware from the top down, and operates it as a proprietary network. This model works, but it's expensive and selective. Companies build where returns are highest, leaving vast swaths of the world underserved.
What if there was another way? What if ordinary people could collectively build global infrastructure by running hardware from their homes, cars, and businesses, coordinated not by a corporate hierarchy but by crypto-economic incentives encoded in a protocol?
This is DePIN: Decentralized Physical Infrastructure Networks. It sits alongside DeFi and NFTs as a third pillar of crypto. Instead of just moving money or ownership records, DePIN pushes blockchains into the physical world. The model replaces centralized capital budgets with distributed participation, turning hobbyists and small businesses into the operators of networks that rival, and sometimes surpass, what incumbents have built.
By late 2022, one such network called Helium had close to one million small wireless devices deployed across nearly 200 countries. Most weren't carrier-owned towers. They were modest boxes plugged into living rooms, shops, and warehouses, installed by individuals who earned cryptocurrency for providing IoT (Internet of Things) coverage. In just a few years, this community-built network grew large enough to rival traditional deployments that took telecom companies decades and billions of dollars to establish.
Around the same time, Hivemapper launched a decentralized mapping network that pays drivers to collect street-level imagery with dashcams. By 2024, contributors had mapped roughly 16 million unique kilometers of roads, more than a quarter of the world's road network, at a pace that outstripped Google Street View's historical expansion. They used ordinary cars and token incentives instead of dedicated mapping fleets.
These aren't quirky side projects. They represent a broader pattern showing how the model works in practice. Most projects rely on the same base crypto infrastructure, but their success is constrained by hardware and local economics, not just code.
Section I: The Infrastructure Gap
Dead zones on your phone, $200 cloud bills, and towns stuck on 3 Mbps DSL are not accidents; they’re features of how we fund infrastructure. For centralized incumbents like AT&T or AWS, every new tower, fiber run, or data center is a spreadsheet exercise in capital expenditure versus long-term return.
Running fiber optics to a rural town of 500 people requires millions of dollars in upfront investment, yet the monthly subscription revenue might take decades to repay that debt. The result is the centralization trap, where infrastructure is deployed only in dense, profitable urban centers, leaving the long tail of the global population underserved.
The DePIN Inversion
DePIN flips this model on its head. Instead of one company raising billions to build a proprietary network, a protocol coordinates thousands of people to deploy hardware themselves. There’s no top-down rollout plan; the network grows wherever the incentives make it worth someone’s time to plug a device in.
The core bet is simple: if you pay people in the right way, a community can spin up a global network faster and cheaper than any single company. The protocol doesn’t lay cables or buy servers; it runs the marketplace, letting anyone plug in hardware and earn, with the economics baked directly into code via native tokens and on-chain metering.
A network with no CEO still needs a way to make decisions. As with other crypto protocols, DePIN networks rely on token-based governance (the mechanisms explored in Chapter XII). Token holders vote on critical parameters like hardware specifications, emission schedules, and protocol upgrades. Often, these voters are the operators themselves, meaning the network is owned and operated by the supply side. This creates a self-organizing system where individual profit motives are harnessed to build a unified public utility.
Section II: The Economic Engine
Not all DePINs ask for the same kind of commitment. Hardware-heavy projects require operators to buy and install dedicated devices, pay for power and internet connectivity, and handle maintenance. Lighter-weight designs instead ask you to share existing resources: your home internet, phone sensors, or spare storage. This shifts the cost from hard cash to your time, attention, and risk tolerance.
If the goal is to build a global network without a central treasury, the immediate hurdle is the cold-start problem. In the early days, a network has no utility and therefore no customers. A ride-sharing app with no drivers has no riders, and a telecom network with no towers has no subscribers.
While a traditional startup burns venture capital to subsidize this phase, DePIN protocols utilize token emissions as a temporary subsidy. Early adopters purchase and deploy hardware not because there is existing demand, but to capture these emissions, effectively acting as angel investors who are paid in equity-like tokens rather than in cash.
The Emission Schedule
This mechanism only works if two things line up: the emission schedule and real demand for the token. If tokens are released too aggressively, inflation crushes their value, and rewards stop mattering. If they're released too slowly, nobody is willing to buy expensive hardware in the first place.
In the early stages, most of the real-world cost is actually carried by passive buyers. Operators sell part of their rewards to cover hardware and operating costs, while speculators absorb that sell pressure. These speculators aren't funding the network out of altruism; they're betting it will be worth much more later. Bitcoin followed this same pattern: miners mined to make money and sell BTC to pay their power bills, while long-term holders financed the build-out by buying what miners had to dump.
DePIN projects like Helium often employ a halving schedule similar to Bitcoin's, where rewards drop at fixed intervals to create urgency. To prevent hardware from clustering in already saturated cities, these schedules frequently include geographic multipliers that offer higher rewards for deploying in underserved areas. These geographic zones discretize the world into reward-bearing cells, steering the physical growth of the network through economic code.
The Transition to Real Revenue
Emissions are a bootstrap subsidy, not a business model. In the long run, a DePIN network must be funded by real customers paying real money. For the token to have durable value, some portion of that external revenue needs to flow back to token holders, whether through a treasury, fee-sharing, or buy-and-burn mechanics.
A common approach is the burn-and-mint equilibrium, similar to buyback-and-burn mechanisms in token economics (Chapter XII). Using the network requires burning the native token, permanently reducing its supply. If those tokens were first bought with outside capital (for example, customers paying in fiat while an intermediary buys and burns tokens on their behalf), then network revenue effectively repurchases and destroys supply. Over time, this can support the token price and keep operator rewards economically meaningful, assuming usage grows large enough relative to new issuance and selling pressure.
Helium demonstrates this design in practice. Access to the network is paid for in Data Credits, non-transferable units with a fixed dollar price per data packet. These credits are created only by burning the native HNT token at the current oracle price. For packets to flow, someone must ultimately spend money, acquire HNT, and burn it into credits. Every packet transmitted corresponds to a small amount of HNT supply being retired.
This structure links HNT's value to actual network usage rather than speculation alone. However, the strength of that link depends on scale. As long as token emissions and speculative trading dominate, burning remains mostly a structural feature waiting for real demand.
A DePIN network becomes economically sustainable only when usage-driven burns grow large enough to meaningfully offset emissions and fund operator rewards without relying on an endless stream of new speculators.
Revenue Model Diversity
Beyond Helium, most DePIN networks experiment with various revenue models that must eventually replace emissions as the primary source of value. Some charge directly for consumption through per-gigabyte storage fees or per-compute-hour costs. Others layer on subscription plans for predictable recurring income, particularly in connectivity and SaaS-like offerings. Many capture transaction fees on in-network payments or monetize aggregated, anonymized datasets sold to enterprises, research institutions, or application developers.
Token Supply Mechanics
Token supply mechanics vary widely across networks, even when they’re solving similar problems.
Take Filecoin, a decentralized storage network where users pay to have their data stored by independent providers. Each transaction on Filecoin includes a base fee that is burned, permanently removing those tokens from circulation, while storage providers earn new tokens through block rewards and user-paid storage deals. This combines inflationary issuance (to reward providers) with built-in deflation (via burned fees).
On the compute side, Render Network coordinates GPU providers to perform rendering and AI workloads. Here, token burns are tied directly to completed jobs: when users pay for rendering, a portion of the tokens involved is destroyed, so supply reduction is explicitly linked to delivered compute.
Many other DePIN designs add a third ingredient: staking. Operators must lock up tokens as collateral to participate, taking those tokens out of effective circulation. If they fail to meet service guarantees or act dishonestly, part of that stake can be slashed (similar to validator slashing in Proof-of-Stake systems, covered in Chapter II). Between burned fees, job-linked burns, and staked collateral, each network assembles a different mix of inflation and deflation to align incentives and anchor long-term token value in actual usage.
When these systems work, they exhibit powerful network effects. Early deployments may see little usage, but as coverage density increases and more applications integrate the network, usage and revenue per node can rise non-linearly. When the cycle runs in reverse, the network faces a potential death spiral: usage stalls, token burns decrease, the price drops, and hardware operators begin unplugging their nodes as rewards no longer cover electricity, bandwidth, and maintenance.
Section III: The Trust Machine
The economic incentives described above are powerful, but they create a secondary problem: fraud. If a network pays anonymous actors to provide wireless coverage or store files, bad actors will inevitably try to claim rewards without doing the work.
Therefore, the technical architecture of DePIN exists primarily to answer one question: did you actually provide the service? This requires an ecosystem of service providers who deploy the physical assets, and validators and oracles who confirm the work.
The Roles and Responsibilities
The verification process relies on distinct, incentivized roles. At the foundation are the service providers, the hardware operators who deploy and maintain the physical assets, such as hotspots, storage servers, or sensors.
Service providers typically must meet minimum hardware, connectivity, and uptime requirements. Performance monitoring tracks throughput, latency, and reliability, and many networks maintain on-chain or off-chain reputation scores that route more traffic and fees to consistently reliable operators.
Monitoring their performance are validators and oracles. Unlike service providers, validators contribute computational resources rather than physical infrastructure, confirming cryptographic evidence of service. In many networks, these validation duties are further specialized: challenge generators create randomized verification tasks, while witness networks independently confirm responses, adding redundancy against collusion.
Oracles serve as the bridge between the physical and digital worlds, verifying off-chain data such as local weather conditions or vehicle location and relaying it on-chain for reward calculations. (Oracle networks and their security considerations were introduced in Chapter VII's infrastructure dependencies section.)
Proof-of-Coverage for Wireless Networks
In wireless networks, verification is achieved through Proof-of-Coverage. Hotspots periodically issue radio challenges, encrypted packets sent over the airwaves, to their neighbors. When a nearby hotspot witnesses this packet, it reports the signal strength to the network, creating a cryptographic map of coverage anchored in physical radio propagation.
A key tuning knob is the challenge frequency. Challenge too often and verification floods the network with data and cost; challenge too rarely and attackers have wide windows to spoof coverage without being caught. Most designs are a compromise: enough probes to make cheating unprofitable, but not so many that verification overwhelms the economic value of the service.
To represent coverage in a way that resists gaming, the world is often divided into a hexagonal grid that prevents tightly co-located hotspots from claiming outsized rewards for overlapping service. To manage the massive data load of verifying millions of radio pings, many implementations offload computation to specialized oracle networks (introduced earlier in this chapter and covered in depth in Chapter VII), which aggregate raw witness data and post summarized proofs to a high-throughput blockchain such as Solana.
Proof-of-Spacetime for Storage Networks
Storage networks face a different challenge: proving that data is not just stored, but persists over time. Filecoin replaces the legal contracts and service-level agreements of centralized providers with two cryptographic systems that target these questions directly: Proof-of-Replication and Proof-of-Spacetime.
Proof-of-Replication is used when the deal starts. The storage provider takes the client's data, creates a uniquely encoded copy of it on their own hardware, and produces a proof that this specific copy exists on their disks. This prevents a dishonest operator from pretending to offer large capacity while actually reusing the same underlying data for many different clients.
Proof-of-Spacetime is used over the lifetime of the deal. The network needs to be convinced that the data is still there, but checking every byte of every file all the time would be too costly. Instead, Filecoin uses probabilistic sampling. On a regular schedule, roughly once per day, the protocol asks the provider to prove they still hold randomly chosen pieces of the stored data. If they can respond correctly and on time, the network treats that as strong evidence that the full encoded copy remains available.
In Filecoin, these ongoing checks are implemented in a system called WindowPoSt, short for Window Proof-of-Spacetime. Each time window comes with its own set of challenges and proofs that the provider must submit. If they miss these proofs or submit invalid ones, part of their locked collateral can be slashed and their future rewards reduced. Reliability is enforced not by courts, but by automatic economic penalties that trigger when the proofs fail.
The Location Problem
The most difficult technical vector to secure is location. Because many networks incentivize geographic expansion, they are vulnerable to GPS spoofing, where operators use software to fake their coordinates and harvest rewards intended for underserved areas. The defense against this is a constant arms race.
Protocols employ triangulation to validate location via signal strength between peer devices, and increasingly use proof-of-location hardware with secure elements that sign GPS data at the chip level. Some networks, like Hivemapper, add a layer of AI-based behavioral analysis, flagging non-organic patterns or using visual data to confirm that a camera is actually moving through the physical world.
Many designs also rely on stake-based deterrence, requiring operators to lock tokens that can be slashed if manipulation is detected, and on community reporting tools that allow participants to flag suspicious deployments for targeted audits. Together, these mechanisms attempt to shift location spoofing from a low-risk software trick into a high-risk economic gamble.
The Building Blocks
These economic and verification primitives are the building blocks of DePIN. Different infrastructure types recombine them in different proportions, trading off hardware cost, proof complexity, regulatory exposure, and demand patterns.
To see how this works in practice, it is useful to examine how DePIN architectures diverge across the three primary challenge domains: geographic coverage, data persistence, and computational resources.
Section IV: Categories and Implementation
With the foundational mechanics established, we can now examine how different DePIN projects apply these principles to solve specific infrastructure challenges. Each category faces unique technical and economic hurdles that shape its architecture.
Geographic Coverage Networks
Geographic coverage networks aim to fill physical space with hardware in order to provide connectivity or environmental data. Their core challenge is incentivizing deployment in the right places and proving that the deployed hardware is genuinely present and active.
Wireless Connectivity: The Helium Model
In wireless connectivity, Helium serves as the archetype for the field-deploy model. By incentivizing individuals to host low-power wireless hotspots designed for IoT devices, the network deployed over 900,000 nodes across roughly 170 countries, validating the thesis that token incentives can finance massive capital expenditures without a centralized carrier balance sheet.
Historically, rewards for operators were split between providing coverage, witnessing other hotspots' challenges, and actually transferring user data, directly translating the emission and burn mechanics discussed earlier into physical network growth.
However, Helium also illustrated the volatility of the model. As the network pivoted toward 5G cellular coverage and Wi-Fi offload in later years, it had to manage the complex transition from a pure coverage-building phase to a usage-generation phase, competing directly with established carriers on reliability and customer experience, while governance debated how quickly to shift rewards from coverage to actual traffic.
Mapping Networks: Hivemapper vs. Google
Mapping and sensor networks confront similar coverage challenges in different data modalities. Hivemapper challenges Google Street View by paying drivers to mount dashcams, resulting in a map that can update weekly rather than annually.
By early 2026, Hivemapper contributors had mapped more than 500 million kilometers of roads and reported coverage of roughly one-third of the global road network, a remarkable pace for a network only a few years old. Google, by contrast, has accumulated on the order of ten million miles of Street View imagery over more than a decade, with near-universal coverage in many countries but slower refresh cycles.
Hivemapper's comparative advantage lies not in absolute coverage, where Google still dominates in many regions, but in freshness and marginal cost. Traditional mapping requires expensive fleets of dedicated cars and staff; DePIN turns data collection into a background task for existing drivers, whose rewards depend on location and novelty of coverage.
Because those rewards are location-sensitive, Hivemapper leans heavily on proof-of-location techniques and AI-based validation to ensure that images correspond to real-world streets rather than fabricated or replayed footage.
Environmental Sensor Networks
Environmental sensor networks extend the same pattern to weather and air quality. WeatherXM operates personal weather stations spread across dozens of countries, cross-validating their readings against nearby stations and satellite imagery and rewarding operators with tokens based on data quality and consistency.
The value proposition is hyper-local coverage that national meteorological agencies or commercial providers would struggle to justify financially. Planetwatch applies a similar playbook to air quality: calibrated sensors, often installed in homes, offices, or street fixtures, feed regulatory-grade measurements into the network.
Token rewards are tied to sensor class and sustained uptime, and the resulting datasets are used for public health research, climate analysis, and, in some cases, regulatory monitoring. Here, the trade-off with centralized systems is not purely cost; traditional networks may offer more controlled and audited instrumentation, but DePIN can produce vastly denser coverage if it solves the problems of calibration, fraud, and long-term operator incentives.
Data Persistence Networks
While geographic coverage networks focus on where hardware is deployed, data persistence networks focus on whether data remains available at the right price over the right time horizon. Their primary challenge is replacing corporate contracts and service guarantees with cryptographic and economic enforcement.
Filecoin: The Open Marketplace
Filecoin operates as an open marketplace for storage. Miners compete to offer capacity, and clients negotiate deals specifying price, duration, redundancy, and geographic preferences.
Public analyses have found that, in some periods, advertised Filecoin storage prices have been on the order of a few dollars per terabyte per year, compared to twenty to thirty dollars per terabyte per year for standard AWS S3 tiers, implying an order-of-magnitude cost difference when promotional incentives and aggressive competition are in play. These headline numbers must be treated cautiously: effective cost depends on replication factors, retrieval pricing, and operational complexity. Nonetheless, the competitive pressure is real.
IPFS, the InterPlanetary File System (also mentioned in Chapter XI's NFT storage discussion), sits alongside Filecoin as the addressing and distribution layer, identifying files by their cryptographic hash rather than their location. Filecoin adds the incentive layer to ensure that critical content persists over time, using Proof-of-Replication, WindowPoSt, and slashing to enforce deals without traditional legal contracts.
Arweave: The Endowment Model
Arweave (also mentioned in Chapter XI's NFT storage discussion) takes a fundamentally different approach, offering permanent storage via an endowment model. Users pay a one-time, upfront fee that is effectively invested into a storage endowment; miners are then rewarded from this pool for storing historical data, ensuring that the network's permanent data archive remains accessible indefinitely without recurring monthly payments.
In practice, this model has produced user-facing prices that have often been in the single-digit dollars per gigabyte range for long-term storage, which can be expensive for very large datasets but attractive for high-value archival content such as cultural artifacts, legal records, or irreplaceable application state.
Here the trade-off with AWS is not only price, but time horizon and control. A corporate provider can change pricing or discontinue a service line; Arweave's promise is that, as long as the network and its token economy survive, the data will remain accessible without further negotiation.
The Storage Stack
Taken together, IPFS, Filecoin, and Arweave illustrate three layers of a DePIN storage stack. IPFS handles how data is addressed and moved, Filecoin offers a market for economically enforced persistence over defined terms, and Arweave provides an option for data that must be preserved for very long horizons or effectively forever.
Each makes different trade-offs between cost, complexity, and assurance.
Computational Resource Networks
The final category of DePIN infrastructure monetizes idle or underutilized processing power rather than physical coverage or storage capacity. Their challenge is transforming a fragmented landscape of heterogeneous machines into something that feels, from the user's perspective, like a coherent cloud.
Render Network: Tapping Idle GPUs
The GPU shortage of 2023 highlighted the inefficiency of centralized clouds, where premium chips were scarce and expensive while consumer-grade GPUs sat idle in gaming PCs worldwide. Render Network taps into this sunk cost.
It aggregates idle GPUs for rendering and AI tasks, dispatching work to nodes that advertise compatible hardware and acceptable pricing. A Proof-of-Render mechanism splits jobs across multiple nodes and verifies outputs via redundancy or cryptographic checks: for example, by re-rendering small portions of a job or comparing hashes of deterministic outputs.
Nodes that return invalid or low-quality results can be penalized or excluded from future work based on reputation.
Akash: General Cloud Computing
Akash extends this model to general cloud computing. It creates a reverse-auction marketplace where tenants specify their requirements (CPU, memory, storage, duration) while providers bid to fulfill them, rather than forcing users to accept the fixed-price menus set by Amazon or Google.
Because many providers have already paid for their hardware for gaming, mining, or existing data center workloads, they can often offer compute at a steep discount compared to centralized cloud margins, particularly for non-critical or burst workloads.
As with other DePIN networks, the theoretical advantage is price and flexibility; the practical constraint is reliability, orchestration complexity, and compliance.
Section V: The Reality Check
Having explored the theoretical models and practical implementations, we must now confront the gap between DePIN's promise and its current reality. Despite rapid deployment metrics and compelling price comparisons, DePIN has not yet displaced traditional infrastructure. The barriers to mass adoption are significant, and they exist in the messy reality of reliability, security, regulation, governance, and user experience.
The Reliability Challenge
Reliability is the most visible hurdle, but it is not as binary as the phrase "five nines" suggests. Enterprise infrastructure demands 99.999 percent uptime for mission-critical workloads, and centralized providers achieve this through over-provisioned capacity, professional operations teams, and contractual service-level agreements.
A DePIN network, by contrast, is an aggregate of thousands of amateur operators with heterogeneous hardware and varying levels of commitment. If a node operator goes on vacation and their internet cuts out, the service degrades.
Protocols attempt to mitigate this through redundancy, reputation systems, and slashing, and for many use cases, personal file backup, non-critical IoT data collection, edge caching for latency-tolerant applications, 99.9 percent uptime at a fraction of the cost may be entirely acceptable.
The challenge is that the highest-value workloads, and thus the largest revenue pools, tend to be the ones that demand the highest reliability. Until DePIN can either directly meet those standards or enable reliable service providers to build on top of it as a wholesale layer, it will be confined to a subset of the potential market.
Security Vulnerabilities
Security sits alongside reliability as a hard constraint. DePIN networks inherit all the usual failure modes of crypto systems: smart contract exploits, misconfigured oracles, attacks on the underlying consensus layer. But they add a new hardware attack surface.
A vulnerability in a popular hotspot model or sensor firmware can be exploited across tens of thousands of nodes at once, undermining the integrity of the proofs the network relies on to decide who gets paid. Protocols can respond with software patches, more aggressive challenge schemes, and slashing, but the combination of standardized hardware and pseudonymous operators makes systemic failures both possible and, in some cases, hard to remediate quickly.
Regulatory Complexity
Regulation further complicates the picture. These networks operate in physical jurisdictions with legal frameworks that are indifferent to blockchain immutability. Telecom networks must navigate spectrum licensing and comply with national telecom regulations. Storage and mapping networks face data sovereignty laws like the EU's GDPR and privacy concerns regarding public surveillance. Environmental sensor networks must align with standards for measurement accuracy and reporting.
A hostile regulatory action, such as banning unlicensed spectrum usage, restricting the export of certain data types, or classifying tokens as unregistered securities, can severely cripple a network's utility. Unlike purely digital DeFi protocols, DePIN projects cannot simply "exit to cyberspace"; their hardware and many of their operators are rooted in specific countries.
Market Cyclicality
All of this sits on top of volatile crypto markets. When token prices crash, rewards can suddenly fall below operators' real-world costs even if usage is steady. This cyclicality makes DePIN unusually sensitive to broader crypto market swings. A bear market can trigger mass unplugging long before the underlying infrastructure thesis is actually disproven.
In practice, many networks experience a brief period of extreme token appreciation early in their lifetime. During this window, headline yields look absurdly attractive in fiat terms, pulling in operators and capital far faster than organic demand justifies. Hardware fleets get built to serve a price signal, not real usage. Expectations anchor around those inflated rewards.
When the inevitable correction comes, often with drawdowns of 90 to 99 percent from peak, the economics flip overnight. The same fleet that looked wildly profitable now struggles to cover electricity, bandwidth, and maintenance. The result is a reflexive unwind: unplugging degrades service quality, which further discourages demand and reinforces the downtrend.
The key long-term metric is whether usage-driven burns can eventually outpace emissions, creating sustainable demand independent of speculation. Most networks never cross this threshold.
The Governance and Usability Paradox
Ultimately, the model faces a governance and usability paradox. Token-weighted governance can be slow, contentious, and captured by large holders, making it difficult to execute the rapid pivots often required in hardware-intensive industries.
Operators, speculators, and end users often have conflicting priorities, and without careful design the decision-making process can amplify short-term interests at the expense of long-term network health. At the same time, the user experience for the demand side (the actual consumers of data, storage, or connectivity) often involves friction-heavy wallets, bridges, and gas fees.
For many potential customers, any theoretical economic advantage is offset by operational complexity and perceived risk.
The Path Forward
Yet it would be premature to dismiss DePIN as merely a proof of concept. The networks in this chapter already demonstrate that loosely coordinated communities can build and operate real infrastructure at scale: global IoT coverage, petabyte-level storage markets, GPU render farms, and live environmental sensor grids.
The question is no longer whether the model works, but where it can scale. Moving from niche experiment to mainstream utility will likely require several key pieces. Networks need standardized service layers that bundle heterogeneous nodes into enterprise-grade offerings. Governance frameworks must move quickly without being captured. Regulatory regimes must learn how to classify and supervise decentralized operators. User interfaces need to make consuming DePIN services feel as simple as using today's clouds.
If those pieces fall into place, DePIN may not replace incumbents outright, but it can reshape them. Traditional platforms could evolve into brands that package, regulate, and resell services built on a global, community-owned substrate of physical infrastructure.