How IPFS is building a new internet from the ground up

Molly Mackinlay on Web3, web apps, content moderation and building an internet that lasts.

IPFS logo in Nyan Cat style.

The IPFS team is building a new kind of internet, memes and NFTs and all.

Photo: Protocol Labs

Molly Mackinlay loves the music app Audius, a decentralized tool that is trying to rethink the way artists own their music and interact with fans. She's a big believer in NFTs, and is looking forward to a world where everything from houses to cars are sold and tracked through the tokens. And she's definitely excited about the metaverse, as long as it's "crazy and open and enables all sorts of creation, which doesn't come from one single company running the metaverse."

In her day job at Protocol Labs (no relation to this publication), Mackinlay spends her time building the infrastructure that will enable all of that. She oversees IPFS, the underlying protocol that could be the future of how data moves between devices, networks and even planets. It's a job that requires wrangling thousands of developers and projects, prioritizing many different ideas about how the future of the internet should work, and trying to convince everyone to jump on board with the decentralization movement.

Mackinlay joined the Source Code podcast to discuss her vision for the future of the internet, what it takes to build an internet that never breaks or crashes and the opportunities Web3 holds for companies new and old, big and small.

You can hear our full conversation on the latest episode of the Source Code podcast, or by clicking on the player above. Below are excerpts from our conversation, edited for length and clarity.

Before you were at Protocol Labs, if I remember right, you were at Google. And then you left Google to come do this wacky, decentralized internet thing. Why? How did you get started working on this stuff?

I started participating in the IPFS open-source community back while I was still at Google. I was on the Google Education team, and I knew some of the folks who were brainstorming up the beginnings of IPFS from college. So I was hearing about this at the same time that I was working on the Google Classroom team, building tools for teachers and students.

I was the PM, designing and building mobile apps and was thinking about things from an offline perspective. Mobile phones, more so than the web product of Classroom, have this ability where you can take them all around with you and use them offline. And so the question was, what could we do from an offline perspective? And in observing classrooms and talking with teachers and students, we kept running up against, well, it'd be really great if a teacher and student who are in the same physical classroom with each other could, say, turn in an assignment the way they currently do with paper, which makes perfect sense. But when you go to the digital world, the way that we've architected the internet and all of these technologies that we depend on every single day, it requires you to sync all of your data to some distant servers somewhere in the ether. And then hopefully, they sync all of that content and sync it back to your teacher.

If your school internet is terrible, which it is almost everywhere in the world, you really struggle with that. And so students were spending like five, 10 minutes of their class time just trying to turn in assignments and sync documents and download videos, and they're all trying to do it at the same time. The exact same document, the exact same video. It's like, this is so broken, the web should work differently.

And that's kind of when I was talking about these problems with some of the folks in the IPFS community. And they're like, "this is exactly our vision for how the internet should work. It should be local-first, it should be peer to peer, you should be able to collaborate with any other person, any other node in the network directly without having to go through an intermediary." It would enable a school with terrible bandwidth to sync a piece of content once and then distribute it to all of your peers and collaborate on it together. That would be a new fundamental way for the internet to work that would just unlock a whole new set of capabilities.

Going to Mars ... you're not going to want to sync every document or every website or every collaborative edit of a Google Doc 14 minutes to fetch content.

And even more futuristic, as you think about us becoming a space-traveling civilization in the next couple of decades, going to Mars, going to the moon, setting up bases there, you're not going to want to sync every document or every website or every collaborative edit of a Google Doc 14 minutes to fetch content. You'd want to be able to do things local-first, you wouldn't want to have to have all of these intermediaries and latencies.

Wait, so IPFS standing for Interplanetary File System is not a joke or tongue-in-cheek. It's for real.

It's an homage to [J.C.R.] Licklider's Intergalactic [Computer] Network, which was an early name for the internet. But it very much is at the core of it. We're not there yet, but when you really stretch it, what would this entail? How is the human race going to evolve? What sort of tools and technologies for this layer that we all depend on, and we build our core capabilities on — how do we want it to evolve? What primitives do we want it to have?

Some of those primitives are user agency, peer-to-peer collaboration and hyper resiliency to all sorts of faults and issues. We can never predict what sort of natural disaster or otherwise is going to disrupt our systems. And so we need to make sure that humanity's information doesn't all get centralized in the Library of Alexandria and then burned. We want it to be hyper-distributed, replicated, resilient across many, many different parts throughout all of the galaxies that humans eventually colonize.

Before we get too deep in this, just give me the basic, kindergarten-level explanation of what IPFS does.

The way I explain IPFS is that it takes the internet model of fetching content and finding data, and instead of addressing everything by its location — what entity in the network is hosting this, facebook.com hosts this picture — it tries to address things by what they are. By the content itself. So a picture, when hosted by anyone, would have the same address, and anyone would be able to fetch it. And they would know that this is the exact picture they were looking for. And they could do that from anyone in the network. They don't have to trust Facebook, who says "yes, this is the picture you were looking for." They can do it from their neighbor, or some random stranger on the internet, and that's all totally fine.

That shift in model unlocks a whole set of new capabilities. The thing that relies on is peer-to-peer addressing and content addressing, so, being able to actually network computers to talk to each other directly, and then enabling them to address the content that they're looking for, by its cryptographic hash, the digital fingerprint of this file.

A great analogy that a friend of mine uses is, imagine that you're trying to find a book. And instead of telling you what the book is, and letting you identify whether or not you've already read it or have it with you or anything like that, they tell you "well, this book is located in the New York Public Library, on the third floor, fourth shelf from the right, top row, three books in." OK, now you have to go there physically in order to identify what the book actually is. You have to travel to New York, you have to wait until the library is open, you have to get a pass, you have to go in. And then you have to look at the book, and you're like, "Oh, crap, I had this in my backpack the whole time." When you use location addressing, which is kind of how the web was initially built, you don't have the capabilities that we wish it had today.

And so at the core of it is switching the web from using location addressing to content addressing, which then unlocks being able to fetch content from anyone. And IPFS is a kind of file system and set of tools built around trying to make that really easy for building new applications. It has a file system-like access, so you can add files and fetch files from different peers in the network, all using this technology.

The thing that comes up the most often when people talk about IPFS is permanence: You can't lose data. You can't censor data. It lives beyond any individual person's control. And I think that's really powerful. Are there other sort of core tenets of what you're trying to do that are on that same level? Or is everything kind of a knock-on effect from that permanence?

To me, "user agency" is actually the term I use more than permanence. Permanence is really important from an addressing perspective, but when you frame it from the lens of user agency, it's really about giving you control over your own experience browsing on the web. That implies permanence, because if you want to have a piece of content on your machine, you want to host it, you want to access it, that's your node's prerogative. But it doesn't necessarily mean that you can force someone else to store content on the internet that they don't want to. And it means that if you don't want to load content that some other node happens to be storing, you don't have to.

It's really about giving you control over your own experience browsing on the web.

Trying to really put the tools and control in the hands of the individual of the node operator is one of the primitives I think is really important. And what that implies around permanence and censorship is that we as a society become more flexible. It leans more into things like freedom of speech, where we're trying to navigate that effectively with each other's peers, and that you as node operator are responsible for the speech you make within the network.

One of the criticisms people have of systems like this is that you can only support them if you're like a free speech purist, who believes everything should be available to everyone, nobody should be able to take anything down. How do you think through that tension in your head, between good governance and taking care of people versus this core belief in free speech?

It's a tricky subject, right? I think there's some really, really valuable work that's being done by groups like the Element team at the Matrix Foundation, where they're working on distributed content moderation schemes, whereby you as an individual or a collection of individuals can help filter your own content. You can subscribe to the set of filter lists you want on the internet. Maybe you hate seeing pictures with the color blue, for some reason, and so you could subscribe to a filter list that's like, "no blue images, blue images make me sad, I don't want any of those!" And you can curate the set of things that fall within that filter list, and your experience of the web is tuned for what you as a node operator wants.

I think where you really don't want to get into is where those things are applied on top of the user, restricting their own way of accessing the internet, which is just how you get into all sorts of really bad forms of content moderation that we've seen play out super poorly in the Web2 world. It's a centralized group that's now put in the position of having to moderate all policy on the internet across every global border. And this is non-trivial, and probably shouldn't be in the hands of, say, YouTube's centralized content moderation team to be making all of those decisions for everyone around the world.

I think it's going to evolve to effectively look like you as a node operator, as someone who's publishing content, can always take that content down. No one can force you to keep content around. And you can absolutely delete content from the network, that's totally possible. But when it comes to any central node being able to black out a highly desired and valuable and useful piece of content from all nodes across the internet, that's also probably not something that we want to put in the hands of one central party. So it's a balance, how we build up the right set of tools that enable groups to self-moderate the content as it relates to their network, without giving them the reins to black out everyone's, you know, Wikipedia page with words and names that they don't want included.

You're making me realize that you basically have to totally rethink what apps and platforms mean and look like. Do you find yourself meeting with people and having to pick up pieces of their brain off the floor, 10 minutes into your conversations?

And pieces of my own brain!

I remember in my first couple months on the IPFS team, sitting down and thinking about, OK, when I first got started in computer science, one of the easy things to do was get started building little games. So a little chess app, a little Sudoku app, how would I do that in a peer-to-peer way over Web3? So, OK, chess is relatively easy. You just alternate between nodes being able to make moves. And our nodes would be connected to each other. But if you went offline, and you wanted to rejoin, how would you find the other node? If you wanted to have a timer, whose version of time would we be using? There's actually an entire blockchain for this now, that gives you a decentralized time-stamping service.

You have to think about how you architect your applications in a different way, when you move into Web3. Back then, if I wanted my own decentralized time-stamping solution, I would have had to come up with my own solution to that problem. Now, there's a really easy API, I could subscribe to that.

You have to think about how you architect your applications in a different way, when you move into Web3.

But that's the layer of additional thinking about the problem and additional infrastructure that needs to be created so that it's really easy for an application developer to lift their brain from the Web2 side of the world and transfer it over to the Web3 world without scattering some pieces. It definitely takes some getting used to.

That's one of the reasons why you see so many of the Web3-native applications growing from scratch within this new area of the world, because it's difficult to transition a massive enterprise over into this new model. I think some groups have been more successful at kind of growing new endeavors, like maybe the Microsoft ION group, where they're growing kind of a new Web3 identity initiative. But they're not trying to just transition an existing Web2 model and presume that all the Web3 primitives are going to be one-to-one with how they would have designed something in Web2. It definitely still takes some work, and the work that we're doing every day is making it more accessible and easier for people to make that transition.

Have you spent time talking to folks at the Googles and Facebooks and Microsofts of the world about what all of this looks like? A lot of people are going to start to ask, "How do I build a $100 billion business out of this?" Which is a complicated question. And other people are going to say, "I already run a $100 billion business. What am I supposed to do with Web3?" Are the answers to that getting better on your end?

Yeah. I mean, the homegrown businesses within Web3 are absolutely scaling. OpenSea is now what, a billion dollars per month or something crazy? There are definitely businesses that are homegrown within Web3 that are quickly scaling to top some of those lists and be very interesting from a traditional Web2 scale perspective.

When the conversation comes to large, existing businesses, the feedback I give is to start understanding the space first. Don't start with, "great, we're going to transition an existing business directly on top of this network." Start with, what are the components of your business that make the most sense?

We've actually been working with a ton of super large institutions who have massive datasets that they want to be hyper resilient against all sorts of faults, and stored cheaply around the world, to utilize things like IPFS and Filecoin for that purpose. Groups like the Internet Archive, the USC Shoah Foundation, the Max Planck Institute and others. Those are big, brand-name institutions who are coming over into the space and seeing the value that it can provide to them, but they're not taking an existing cookie-cutter application and not rethinking some of the kind of interface points, or how to make the best use of this technology for their purpose.

Personally, I think that the pathway for big businesses into Web3 will be one that charts the incremental value that they get at each point. The first step, working with someone like Netflix, is looking at their build systems, and how you offer incremental value in decentralizing a build system using content addressing. Great, that makes incremental sense for a piece of the business. And then you go from there step by step until voila, you have Web3-ized all of the core pieces that the entire business is built upon.

To the point about getting more people on board, it seems like the next key phase of this is to get beyond the enthusiasts, who are willing to do the work and are interested in this on its own merits. There's this big group of people who think this stuff is interesting … because they think it's interesting. And then there's a much bigger group of people who just absolutely could not care any less. And that's ultimately the group of people you have to get, right? They just want to check their email, or listen to music or find a document. How are we doing it at starting to move toward some of those people?

I would personally classify myself as someone who does not fall into the "thing, but decentralized" camp. The reason I got excited about it in the first place was the capabilities it offered. I want teachers to be able to return assignments to students in the same classroom. That's not "a thing, but decentralized," that's a new capability that should be part of the web. And we should just make it freaking work.

That's not "a thing, but decentralized," that's a new capability that should be part of the web. And we should just make it freaking work.

A lot of folks who are in the ecosystem already and coming to the ecosystem today, they do see new capabilities that are coming out of the decentralization layer that weren't there before. And they're coming for those capabilities, and for the opportunities and businesses that can be built bringing those opportunities to the masses.

A lot of the groups who are building applications in the Filecoin/IPFS space today, like they are leaning into what is possible. It's not just "x but decentralized," it's "x but no one can intermediate the music creator from their audience," or "we can collaborate offline on our Google Doc without having any other node," or "we can store massive amounts of video data for teachers who are teaching in India."

Those are capabilities, and people who see how to compose these new primitives into unlocking those capabilities are being successful in the Web3 space today. And I think folks who got into it in the early days are being augmented by a much larger cohort of people who are getting into it to demonstrate how that makes life better for everyone. So yeah, I think that wave has already begun. And I'm really excited to see it continue.

We're not yet at the late majority of folks who are like, "I don't want to have to do any work, and I don't want to have to invent the future, just give me the thing that already works." We're still in the builder phase. But it's not just the decentralization, it's builders who want to change the world.

Theranos’ investor pitches go on trial

Prosecutors in the Elizabeth Holmes fraud case are now highlighting allegations the company sought to mislead investors.

The fresh details of unproven claims made about the viability of Theranos' blood tests and efforts to conceal errors when demonstrating testing equipment added to the evidence against Holmes, who is accused of fraud in her role leading the company.

Photo: David Paul Morris/Bloomberg via Getty Images

The Theranos trial continued this week with testimony from Daniel Edlin, a former product manager at the blood-testing startup, and Shane Weber, a scientist from Pfizer. Their testimonies appeared to bolster the government's argument that Holmes intentionally defrauded investors and patients.

The fresh details about audacious and unproven claims made about the viability of Theranos' blood tests and efforts to conceal errors when demonstrating testing equipment added to the evidence against Holmes, who is accused of fraud in her role leading the company.

Keep Reading Show less
Aisha Counts

Aisha Counts (@aishacounts) is a reporting fellow at Protocol, based out of Los Angeles. Previously, she worked for Ernst & Young, where she researched and wrote about the future of work, emerging technologies and startups. She is a graduate of the University of Southern California, where she studied business and philosophy. She can be reached at acounts@protocol.com.

The way we work has fundamentally changed. COVID-19 upended business dealings and office work processes, putting into hyperdrive a move towards digital collaboration platforms that allow teams to streamline processes and communicate from anywhere. According to the International Data Corporation, the revenue for worldwide collaboration applications increased 32.9 percent from 2019 to 2020, reaching $22.6 billion; it's expected to become a $50.7 billion industry by 2025.

"While consumers and early adopter businesses had widely embraced collaborative applications prior to the pandemic, the market saw five years' worth of new users in the first six months of 2020," said Wayne Kurtzman, research director of social and collaboration at IDC. "This has cemented collaboration, at least to some extent, for every business, large and small."

Keep Reading Show less
Kate Silver

Kate Silver is an award-winning reporter and editor with 15-plus years of journalism experience. Based in Chicago, she specializes in feature and business reporting. Kate's reporting has appeared in the Washington Post, The Chicago Tribune, The Atlantic's CityLab, Atlas Obscura, The Telegraph and many other outlets.

Protocol | Policy

8 takeaways from states’ new filing against Google

New details have been unsealed in the states' antitrust suit against Google for anticompetitive behavior in the ads market.

Google is facing complaints by government competition enforcers on several fronts.

Photo: Drew Angerer/Getty Images

Up to 22%: That's the fee Google charges publishers for sales on its online ad exchanges, according to newly unredacted details in a complaint by several state attorneys general.

The figure is just one of the many details that a court allowed the states to unveil Friday. Many had more or less remained secrets inside Google and the online publishing industry, even through prior legal complaints and eager public interest.

Keep Reading Show less
Ben Brody

Ben Brody (@ BenBrodyDC) is a senior reporter at Protocol focusing on how Congress, courts and agencies affect the online world we live in. He formerly covered tech policy and lobbying (including antitrust, Section 230 and privacy) at Bloomberg News, where he previously reported on the influence industry, government ethics and the 2016 presidential election. Before that, Ben covered business news at CNNMoney and AdAge, and all manner of stories in and around New York. He still loves appearing on the New York news radio he grew up with.

Protocol | Workplace

This tech founder uses a converted Sprinter van as an office on wheels

The CEO of productivity startup Rock likes to work on the road. Here's how he does it — starting with three different WiFi hotspots.

Kenzo Fong, founder and CEO of the 20-person productivity software startup Rock, has been working out of his converted Mercedes-Benz Sprinter van since the pandemic began.

Photo: Kenzo Fong/Rock

Plenty of techies have started companies in garages. Try running a startup from a van.

In San Francisco, one software company founder has been using a converted Mercedes-Benz Sprinter van — picture an Amazon delivery vehicle — as a mobile office.

Keep Reading Show less
Allison Levitsky
Allison Levitsky is a reporter at Protocol covering workplace issues in tech. She previously covered big tech companies and the tech workforce for the Silicon Valley Business Journal. Allison grew up in the Bay Area and graduated from UC Berkeley.
Protocol | Policy

Most Americans want AI regulation — and they want it yesterday

In a poll, people said they wanted to see artificial intelligence technologies develop in the U.S. — alongside rules governing their use.

U.S. lawmakers have only just begun the long process of regulating the use of AI.

Photo: Louis Velazquez/Unsplash

Nearly two-thirds of Americans want the U.S to regulate the development and use of artificial intelligence in the next year or sooner — with half saying that regulation should have begun yesterday, according to a Morning Consult poll. Another 13% say that regulation should start in the next year.

"You can thread this together," Austin Carson, founder of new nonprofit group SeedAI and former government relations lead for Nvidia, said in an email. "Half or more Americans want to address all of these things, split pretty evenly along ideological lines."

Keep Reading Show less
Ben Brody

Ben Brody (@ BenBrodyDC) is a senior reporter at Protocol focusing on how Congress, courts and agencies affect the online world we live in. He formerly covered tech policy and lobbying (including antitrust, Section 230 and privacy) at Bloomberg News, where he previously reported on the influence industry, government ethics and the 2016 presidential election. Before that, Ben covered business news at CNNMoney and AdAge, and all manner of stories in and around New York. He still loves appearing on the New York news radio he grew up with.

ai
Latest Stories