Artificial Ignorance

Charlie Guo

Interviews with founders, investors, and deep thinkers on artificial intelligence and its impact on our world. For more deep dives and news stories, visit www.ignorance.ai. www.ignorance.ai

Episodes

  1. 08/07/2024

    How a $3000/hour escort uses AI to automate sex work

    As someone who grew up with the Internet, I’ve occasionally found myself in strange corners of it. But more often than not, I’ve walked away from those experiences with more knowledge than before, and a constant reminder to stay open-minded. So when I stumbled upon Adelyn Moore, an escort who charges $3000 per hour and tweets about using AI for sex work, I wanted to learn more. In our candid and thought-provoking interview, Adelyn - a self-described "autistic courtesan" - offers a fascinating glimpse into the world of modern sex work and its intersection with technology. As the adult industry grapples with the rise of AI-generated content and "digital girlfriends," Adelyn sheds light on the complex interplay between authenticity, technology, and human connection. Her experiences challenge common assumptions about sex work and reveal how the world’s oldest profession can leverage its newest technology. “When choosing a profession, I took the Lindy Effect literally.” – Adelyn Moore Three key takeaways AI-generated adult content still struggles with realism. While there are attempts to create ever more realistic AI porn, the current technology still struggles to produce convincingly realistic images and videos. The skin doesn't look real. The video quality at AI right now is mediocre. Image generation always looks weird - it looks like hentai often. It's not like at a point right now where it's good enough in the ways that you can actually be like utilized. Likewise, imperfection and "realness" are increasingly valued in the adult industry. As “perfect” photos and videos become more prevalent, there's a growing appreciation for content that shows authenticity and flaws. A lot of times I'll just disregard [online] photos because I'm like, “Okay, this photo just looks too manicured.” It's just boring. There's this heightened level of sensitivity because people see so little of just being messy. I have a video on my OnlyFans where I'm awkwardly… like, I wish I could do a sexy thing where I like rip open a condom with my teeth and I'm just trying and I'm like, I can't. I don't know if that's particularly sexy. But I like that stuff. The escort-client relationship offers a unique form of intimacy that may be difficult to automate away. The controlled environment and strict boundaries of these interactions allow for a rare level of openness and vulnerability than in other social contexts. When I first started [escorting], I thought it was always interesting how like people would tell me stuff that like, they were honest, that they didn't tell anyone else in their lives. … People often have a certain amount of vulnerability with me immediately. It's also this weird thing where it's, you have this parasocial relationship because they've met me and they've often people will be like, “I've been following you for like a year. I've been following you on Twitter for like months,” or like the fact that people will be like “I created a special anon account just to follow you.” And it's just that you immediately have this degree of intimacy that I have not seen replicated in any other part of my life. And three things you might not know: * Escorts hang out on social media just like everyone else; that thirst trap you’re scrolling past might actually be content marketing. * While most text-to-image generators like Midjourney have strict filters on nudity and adult content, other projects are working diligently to enable fully uncensored AI images. * Sex workers can perform pretty thorough background checks and screening on potential clients - including requiring IDs, phone numbers, and LinkedIn profiles. Artificial Ignorance is reader-supported. If you found this interesting or insightful, consider becoming a free or paid subscriber. This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.ignorance.ai/subscribe

    32 min
  2. 06/27/2024

    Saving the world with AI and government grants

    At OpenAI's DevDay, I had the pleasure of meeting Helena Merk, the CEO and founder behind Streamline Climate. Streamline is rethinking the way climate tech startups secure vital government funding, but leveraging AI to navigate, automate, and optimize the labyrinthine grant application process. In our interview from March (before the release of GPT-4o or Claude 3.5 Sonnet), Helena shares insights on how Streamline Climate is not just saving time, but potentially accelerating our global response to climate change. From tackling bureaucratic hurdles to leveling the playing field for innovators, discover how AI is reshaping the landscape of climate tech funding and possibly our planet's future. Four key takeaways Bureaucracy is a silent killer. In Helena's experience, many climate grants are ineligible right out of the gate because of incorrect clerical errors or misunderstandings of the grant's eligibility criteria. Ultimately, this wasted time and effort compounds across an entire industry. Half of the reason you would get rejected is because of a clerical error on submitting your files correctly, or filing for something for which you're not eligible, the grant was never meant for you. So people are spending, on average, over 100 hours on a single grant application. And half of those have zero chance of ever winning. Prompt engineering wins out over fine-tuning. OpenAI's advice to Streamline (which mirrors what I've seen elsewhere) is that prompt engineering can get you much farther than you may think. Most people believe they need fine-tuned models, but they really want things to "just work." We've thought a lot about the trade offs and performance benefits of, training on models, fine tuning, etc. And what we've kind of come up with, and this was definitely also part of the advice from OpenAI, is that prompt engineering would get us there fastest in the cheapest way. And probably perform just as good, if not better than training our own models. Besides automating grant writing, there are many potential optimizations in the climate tech space. I learned a ton about the ecosystem around government grants, and we discussed some of Streamline's roadmap beyond just "AI for grant writing." You don't get the money when you win a grant: you have to report on your progress, file all your receipts. After you spend money, you have to get reimbursed. What this means startups is that they have to go get a loan. And that's silly. There's no way the government is going to change this process because of their own risk, but it means that every single company who's winning these grants needs to go get a working capital loan. There's people who provide, specific financing vehicles for this, and we can easily play matchmaker [between grant recipients and financing companies]. Mission-driven founders have to strike a balance between mission and monetization. Helena and I talked about balancing prioritizing the mission with dealing with the incentives inherent in taking VC funding and pursuing growth as a for-profit startup. I've heard people refer to this as missionary versus mercenary type founders. Missionary founders, they're obsessed with whatever mission they're on and they don't really care about anything else. Mercenary founders are really just driven by "I'm going to build a bit of business that like, It goes to the moon." And that is how most of, I think, the Bay area operates. And it leads to pretty decent business outcomes in one way. But I realized that that is never going to be enough for me. I would so much rather run out of money and keep grinding on something that I care deeply about t han pivot into something that's going to be profit generating and maximize returns. I think I learned that when I was working on my first company, Glimpse, where we were working on a video chat company during the pandemic. A pretty safe bet. And I care very deeply about what we started on, which was helping to connect communities of people having deep conversations. It then turned into helping connect remote teams. And the further we got, the more transactional it felt. And I found it really hard. Artificial Ignorance is reader-supported. If you found this interesting or insightful, consider becoming a free or paid subscriber. This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.ignorance.ai/subscribe

    37 min
  3. Getting to the top of the GPT Store (and building an AI-native search engine, too)

    04/24/2024

    Getting to the top of the GPT Store (and building an AI-native search engine, too)

    Back in January, I wrote about what business models might make sense with OpenAI’s GPT Store: As part of that, I included a screenshot of the featured GPTs, which included Consensus, the number one “Research & Analysis” GPT since day one. That said, the company existed well before the GPT Store - they’ve been working for years on an AI-native search engine for research papers. They’re also a part of the latest AI Grant cohort. So, I was pleasantly surprised to learn that one of Consensus's founders, Christian Salem, was a reader! I was equally impressed with our conversation - while I’ve included some key takeaways below, it’s a pretty wide-ranging interview covering RAG and search engine architecture, the value of the GPT Store, how the NFL thinks about product management (and might use AI in the future), and more. Three key takeaways Vector search is not a silver bullet. Rather, it’s another tool that engineers can use as they build search infrastructure. It enables some new capabilities, but it also comes with tradeoffs. Search engines are so much more than just semantic similarity. The vector databases, they're amazing - some of these encoding models are getting better and better. But there's so much that goes into finding relevant documents that isn't just about the distance between two points in vector space. There's things like phrase matching, applying filters, finding synonyms. Sometimes we suggest queries to users. There's fuzzy matching - if you screw up and did a typo, it's kind of weird, but a lot of the vector and encoding models are not as good at typos and fuzzy matching can play tricks on them. There’s real value in the GPT Store. I was surprised to discover that the GPT Store has driven many new users for Consensus, and they currently have ~50% higher retention than other channels. That said, Consensus can capture some value by converting users to paid subscriptions, which isn’t true of free services. So actually we get some of our best users from ChatGPT, which is not something that I would have predicted when we set out doing this. But I think the last time I looked, our day 30 retention is, I want to say like 50 percent higher for users who we acquired through ChatGPT than every other channel. … It's actually turned into awesome users for us who not only use us in ChatGPT, but then come back to our website and use the web application day after day, and many of those users have converted to our premium subscription. So, so far it's resulted in a ton of value. LLMs are not one size fits all. Christian’s comments on using multiple models were similar to the ones from my interview with Andrew Lee - to build a fast, efficient system, you’ll likely need different LLMs for different use cases. There are so many tradeoffs around speed, cost, and quality that it’s hard for one model to win at all three. We were just counting this out the other day, 15 features powered by LLMs are in the product. Only three of them use OpenAI models. The rest all use open source models that we hand fine tune. … When we're assessing which models to use for a new feature or a new task within the product, I think there's a few really important criteria to go through. One, how similar is the task to OpenAI training data? If the task is, “hey, take a bunch of text and summarize it,” GPT-4 is so good at that task. It's seen that over and over and over again. So many users have asked it to do that in ChatGPT. And so they have RLHF on that. And for like a super basic summarization task that is very similar to some core GPT behavior. … Another thing that you obviously have to look at is cost. So, when you ask a question in Consensus for the top 20 papers that we return, we always do a RAG answer from the paper relevant to your question. That is very similar to something that GPT-4 could probably do pretty well. However, it would not be, economical to make 20 OpenAI calls on every single search for both premium and free users. And three things you might not know: * There still isn’t a great way for LLMs to parse structured PDF data, especially in table format. In the case of Consensus, it’s still a human-powered task. * The GPT Store has started testing monetization with a handful of US-based creators but is not yet broadly available. * While the NFL may look like any other company, its "shareholders" are the 32 team owners, meaning new product launches often have to get the approval of owners and their friends and family. Artificial Ignorance is reader-supported. If you found this interesting or insightful, consider becoming a free or paid subscriber. This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.ignorance.ai/subscribe

    41 min
  4. 04/10/2024

    How Intercom is transforming customer support with AI

    In our post-ChatGPT world, few companies have moved to integrate generative AI as quickly as Intercom. Mere weeks after ChatGPT first went viral, Intercom had its AI-powered chatbots in the hands of beta testers, and it's continued to iterate on its products since. That's why I was eager to talk with Fergal Reid, the VP of AI at Intercom and one of the company's key champions of generative AI. Before leading AI at Intercom, Fergal co-founded an ML startup, Synference, which was acquired by Optimizely. We had a chance to talk about Intercom's history building with AI, and how this new technology shift is going to impact customer support reps (for better or worse). In addition to discussing the future of customer support, we discussed Fin AI Copilot, a new AI product from Intercom that's launching today. Like Intercom's existing AI, Fin AI Copilot can pull from internal and external documents, knowledge bases, and other content. But it can also learn from and reference past conversations, and has been built to improve agent interactions, not replace them. Three key takeaways Intercom is expanding from chatbots that talk to customers to copilots that augment agents. The company's first AI product, Fin (now Fin AI Agent), was designed to chat with users about your knowledge base before escalating to human support staff. Now, Intercom is bringing AI capabilities to the support staff themselves, with new features that can write answers, reference past conversations, and supercharge support staff. This framing of a copilot for support really stuck with us. It's a tool. It's got to empower you. It's got to make you faster, more efficient, more effective. And we have for years been really deeply looking at what's the job a support agent does. So often a support agent, especially if they're new or maybe they're getting a question they haven't gotten before, they look at question that comes in and they're like, I don't know what to do here. And then they have to go and read the company documentation, or they have to go and they have to search, they have to go into intercom, they have to search and they have to find the time their more experienced colleague answered this question a week or two before. And we were like, can we use an LLM to do an end run around that? Can we use LLM to just make that experience really seamless? The earliest adopters of generative AI are companies that were both already experimenting with machine learning, and had executive buy-in to move quickly. At Intercom, Fergal's team was able to deploy the first ChatGPT-enabled features within 7 weeks of ChatGPT's launch. That wasn't an accident - they had been playing with this technology for years, laid the groundwork with the rest of the organization. When ChatGPT came out, that was at the end of November and we had features live with our own internal CS team by the holidays that year. ... I'm lucky that we have a very experienced team of some really great folks here who I guess had been in the space for a while. And then we just, we had the executive support we needed you know, because we had done the groundwork because we had talked a lot about how, "Hey, we think to something disruptive here," but we could get the alignment we needed internally to just force velocity of a project like that. Bringing AI to customer support could create more jobs, not less. We've discussed this point in the context of software engineers, but my conversation with Fergal echoed a lot of the same points when it comes to customer support. When each CS rep is much more valuable as a result of AI, does that result in more reps or less? We don't yet have a clear answer. While we don't know how this plays out, one thing we're really confident of at Intercom is that, customer support Is not close to servicing all the demand. There is huge latent demand. Like we see this in every industry study we do. Every time we talk to end users, they want customer support to be dramatically better, faster, friendlier than it currently is. So many customer support experiences are really terrible. And so there's this absolutely latent demand for drastically more customer support. I have to believe that for many businesses, if it becomes way cheaper per customer support rep to deliver great customer support, there's tons more demand there to service. And three things you might not know: * Intercom's AI chatbot, Fin, has a 42% resolution rate on average (and up to 80% resolution rate in the best cases). * When ChatGPT first launched (and even when GPT-4 launched), RAG wasn't a thing for the very first companies trying to build with it - they were inventing the techniques from scratch less than a year ago. * Klarna, a “buy now pay later” company, has said their AI customer support agent does the work of 700 full-time employees. This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.ignorance.ai/subscribe

    36 min
  5. 02/26/2024

    Bridging AI and human creativity

    Something that I’m often thinking about is AI’s ongoing impact on the arts. Clearly, Midjourney and Stable Diffusion have unlocked a new engine for creativity, but it’s just that: an engine. Most of us wouldn’t get much value out of a V8 if it was just dropped in our garage, and most professionals probably can’t go from diffusion model to productive workflow without some extra steps. So designers, especially UX and Figma designers, are still safe from AI for the time being. But there is a lot of change on the horizon - and one of the best people to discuss that change is Harrison Telyan, the co-founder of NUMI, which offers startups access a guild of vetted, professional designers for a flat monthly subscription. Before founding NUMI, Harrison was the founding designer of Imgur, and graduated from the RISD - the Rhode Island School of Design, a world-class design program. Harrison and I talked about his experience rapidly scaling a prior business in Africa, how AI is eating the design world (and the jobs at risk of being eaten), NUMI’s unique, engineering-esque approach to providing a design service, and much more. Three key takeaways Real feedback comes from paying customers. In Harrison’s experience, founders can be reluctant to reach out and talk to their customers directly - and sometimes are even reluctant to charge customers at all. [Something] that I see a lot in founders is how unwilling, maybe not even unwilling, but they have forgotten to actually start the business at some point. I always recommend you chop up your customers in half and start charging them - you will see very quickly the type of feedback that you'll get when you try to separate someone from their money. That's when the real feedback comes. AI has a ways to go before replacing talented designers. Harrison is bullish about AI’s impact on the design community - but he also admits that areas like entry level graphic design work (as opposed to higher level brand identity or UX work), is going to be at risk from AI pretty soon. The real problem that I see though, is none of these [AI] companies have design leaders behind their prompting or their code, and so naturally they're capped. … I'm looking at the landscape and I'm quite bullish on how AI is going to serve the design community. We hear all the time from Guild members at NUMI, is AI gonna replace me? No. It's just gonna allow you to do work faster, more efficiently and you know, it's gonna take away the kinda like rote administrative stuff of design. Not all design agencies are the same. At first, it’s easy to think of NUMI as just another “agency.” But Harrison pushes back on that label - first, because they think of their design community as a guild, not as independent contracts, and second, because they’re building tools and education for the guild to get better, rather than subcontracting work. We always cringe at the word agency when someone's describing us because on the surface, call us whatever you want, but we know what we are. And what we are is a company that was started by designers for designers. And that may not mean much, but when you look at our competition, all of them were started by people in marketing, and then they just create these commodified versions of us that ask for the lowest price at the highest quality with the most communication. We just take a different approach, and that approach is: how can you lift up the designer through technology? How can you remove all the BS from the admin side of what they have to do so that they can get back to designing? It comes down to leveraging tech to remove the BS, to make the designer move faster and put them up on a pedestal. It's actually very similar to how Airbnb thinks about its hosts. Put them up on a pedestal and the rest will work itself out. And that's what we do. And three things I learned for the first time: * Boda bodas are bicycles and motorcycle taxis commonly found in East Africa. * Figma plugins suffer from bit rot - they need to be regularly maintained to keep up with the underlying platform changes. * Many founders seek design services too early, when they really need to be experimenting and talking to customers as much as possible. This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.ignorance.ai/subscribe

    40 min
  6. 01/25/2024

    Funding a new generation of AI companies

    When I went through YC (many moons ago), I still remember hearing stories about startups with incredible traction and growth rates. The YC partners would use them as examples of companies that had found the elusive "product-market fit" and would advise cohorts of startups to follow in their footsteps. One of the companies was Teespring - at the time, it was the fastest-growing e-commerce company ever. Which is why I was excited to talk to Evan Stites-Clayton, the co-founder and CTO of Teespring, who now advises his own cohorts of startups as a partner at HF0. HF0 is a 12-week residency program for technical founders, with a recent focus on AI. They've been profiled in the New York Times, and have already had some impressive companies graduate from the program. They also have incredibly generous terms - applicants give up 3% of their companies for $500K in investment (in contrast to YCombinator's 7% take). Evan and I talked about his early days at Teespring, his transition into AI, and some of the projects he's most excited about - including companies like Krea, iKHOR, and Recursal. Three key takeaways Finding distribution channels is a key part of hypergrowth. That said, it's unclear exactly how AI will impact distribution - while it offers a significant technology shift, how can companies use it to get in front of more users? I think that's also a part of what makes a lot of really great companies tick is that not only do they have the determination and grit and they build a product that people actually want, but they're also riding some kind of wave of, like for us it was Facebook. It was Facebook as a distribution channel, whether it was Facebook groups or then later Facebook ads. But that was the thing that led us to be able to get to that hypergrowth. and I think a lot of companies right now are trying to ride like an AI wave, but it's also important to ask yourself, well what is the AI wave? With Facebook, it was actually pretty obvious - it was this new distribution channel. It was a new way to get into people's feeds. It was a new way to get, an ad in front of someone basically. Now with AI, what is that? Lifelike assistants are an exciting (and underdeveloped) aspect of AI. One of Evan's big interests in AI is digital assistants - he believes it's a major upgrade for consumers to have different personas and friends available 24/7. What I'm most fascinated by is ... a digital friend or a digital assistant that's truly in my life in a way that a human would otherwise be. That's where I think massive value gets unlocked because you enter this world where everybody is able to have all of these different roles in their life. The coach or the helper or whatever it is. But I don't think we're there yet. I think there are a lot of fundamental advances that still need to happen. And it's sort of a UX paradigm of asking ourselves, what is the UX of human interaction that actually creates compelling exchanges? A lot of people think you just can't do that with AI - and I don't agree with that. AI startups need to lean into their agility. This is a sentiment echoed by Andrew Lee of Shortwave - startups have a pretty big advantage in how quickly they can move. Even though incumbents will eventually try to copy them, agile startups will already be on the next frontier of AI features. You have the advantage of having none of that baggage and just being able to completely run in a new direction, a new paradigm. I honestly think the best thing for most early stage founders to do when it comes to thinking about "What if Google builds this? What if Adobe builds this?" is to just ignore those fears and just continue to forge ahead and continue to get users, and continue to listen to the users and continue to build the thing that resonates. ... So ideally you're a position where, let's say you're Krea, you build your AI image generation suite, and then Adobe comes out with its stuff. And by the time Adobe's coming out with its stuff, you're already moving on to the next thing that you're about to release. And so you continue to stay ahead. You continue to stay one step ahead, and then Adobe's gonna release all that stuff they're gonna have to support that. And so you need to keep forging ahead and building more new stuff that's going to continue to keep you relevant. And I think it's very possible for startups to do that because they have less users. They have scrappier dev teams usually, and there's just so much less overhead to like shipping a feature If you are a ten person engineering team than if you are a Google or an Adobe. And three things I learned for the first time: * By 2015, Teespring was doing a million dollars a day in sales. * In contrast to the Transformer architecture that's dominated modern generative AI models, there's an alternative approach called RWKV enabling other types of AI - with different pros and cons. * HF0 originally started as a general-purpose program that changed locations with every cohort. That changed when the pandemic struck (and when Evan hunkered down in their house for months). Artificial Ignorance is reader-supported. If you found this interesting or insightful, consider becoming a free or paid subscriber. This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.ignorance.ai/subscribe

    44 min
  7. 10/18/2023

    The AI email startup that's taking on Gmail

    A few weeks ago, I was lucky enough to sit down with Andrew Lee, the CEO of Shortwave and the cofounder of Firebase. If you're not familiar with Firebase, it's a platform for developers who don't want to host their own backend infrastructure. After being acquired by Google in 2014, it's now a key part of Google Cloud Platform and is used by millions of businesses. These days, Andrew works on Shortwave, an email client that started as a replacement for Google Inbox, but has quickly become a leader in the AI-for-email space. The Shortwave AI assistant can summarize threads, search your inbox, or write your replies for you. Full disclosure: I've been using Shortwave for a couple of months now, and I'm a pretty big fan. I used to use Google Inbox pretty heavily, so it was refreshing to find a worthy successor. Having the ability to summarize emails was the cherry on top - I wanted that feature in Gmail so badly, I made my own Chrome extension to do it. But Shortwave's approach is much, much more thorough than my slapped-together approach. In our conversation, Andrew and I dove into the company's AI architecture, what he learned building Firebase and how he's applying that to Shortwave, and how he thinks about competing with Google/Gmail. Five key takeaways With AI, being lean is a big advantage. Shortwave is able to outpace Google because it can iterate faster with new AI technology, and it doesn't have to worry about working with potential “competitors.” I think we have a few advantages. One is we're just a startup. We can move really fast. So we have something live that works today. Google has Duet AI, which hasn't launched anything at this level. It has some very basic writing features, but most of the stuff we've talked to salespeople about is "coming next year." I used to work at Google. I have some good insight into why it's hard for a big company to move very quickly. People at Google are very sharp and they're good at what they do, but it is a very big challenge to move a huge organization with billions of people forward at a rapid clip. And so we can outrun them. We also have the benefit of being able to use the best technology, wherever it is. I think Google is gonna be extremely reluctant to just start calling the OpenAI API, for example. I think they're be very reluctant to use like open source models from Microsoft, which we do. Making a fast, capable AI app takes more than "just" a few LLM calls. Shortwave's architecture, which they recently detailed in a great blog post, shows the lengths that they've gone to build something that is both more capable than a basic ChatGPT integration, while also being lightning fast. Every time you make a request in our product, there's a dozen LLM calls going out and there's a cross encoder that's running, and there's embeddings that are being generated. The first thing we do is we take your query and the history that you've had with the chat and a bunch of contextual information about the state of the world. For example, what's open on your screen, whether you've written a draft, what time zone you're in, things like that. All so we can figure out what you're talking about, and we ask an LLM "what information would be useful in answering this question?" Do we check your calendar? Do we search email history? Do we pull in documentation? Do we look at your settings? Do we, take the current thread and just stick it in there? There's a whole bunch of stuff that we can do. And once we've determined that, it allows us to kind of modularize our system where we say, "Hey we know we need these three things." And each one of those pieces of information can then go off in parallel and load information. The most interesting one by far is our AI search infrastructure, where we go off and we use a bunch of AI tech to find emails that are relevant to the query and allow it to answer questions that are answerable in your email history. But then we take the output of all those tools, we bring them back together, we throw them in a big master prompt and we have it answer your question. And we do that whole thing, the dozen or so calls, the cross encoder, and the embeddings, and all of that - in three seconds. The current RAG approaches have significant limitations. RAG (retrieval augmented generation) is currently the most popular way of giving LLMs "long-term memory," by fetching relevant documents and handing them to a prompt. But Andrew discussed why that doesn't work amazingly well, and how they're trying to work around it. The kind of standard approach to document search that AI folks are doing is the embedding plus vector database approach. Where you take all of the history, you embed it, you store that in a vector database somewhere, and then you take the query, you embed that, you do a search with cosine similarity, you find the relevant documents, you pull those in and you put them in a prompt. But it doesn't actually produce as good of results as you might like because it only works in cases where the documents that answer your question have semantic similarity with the question itself. Sometimes this is true, right? But if I say, "when am I leaving on Friday," and what you're really looking for is the time of your flight, and that email doesn't have the word "leaving" in there at all. So we wanted to go a step further and say, okay, we want to be even smarter than this. We wanna find things that don't have necessarily semantic similarity. And still answer your question, pull those in. So the way we do that is, we have a whole bunch of different what we call fetchers, a whole bunch of different methods for loading those emails into memory. So we look for names. We look for time ranges, we look for labels, we look for keywords. There's a few other things that we look for. And then we go and we just load all of those emails. We're going to pull all the things that are semantically similar, and the ones that match relevant keywords and the ones in the date range, and the ones involving the right people, et cetera. As always, talking to users is incredibly important. This is one of the things that YCombinator drills into its founders, and with good reason. Shortwave spent over a year experimenting with crazy collaboration features, but ultimately came back to focus on a great single-user experience. When I started this company, I said to myself, I'm not going to be like all those other second time founders that think they know everything. That jump in and think it's going to be easy. I'm going to do this from first principles, and we're gonna talk to our users and we're going to iterate really fast, and we're going to be scrappy. And we did that. We talked to our users, we were disciplined. But it was still just a brutal refresher on how much you have to do that. Like how much you have to talk to users, how much you have to be willing to admit your ignorance and throw out stuff that isn't working. We tried all kinds of features. Like the current state of the product is iteration number, I don't know, 10 or something. For the first year of the product, basically everybody churned. Because we had this much more crazy rethink about how email works, which in retrospect was not a particularly good idea. Sometimes backwards compatibility is inevitable. Many founders are trying to build new and better software, and as a result ship their minimum viable product (MVP) with a bare bones set of features. But sometimes you need to actually support the entire universe of features that your customers actually want - especially if you have established competitors. One of the decisions I wish I would've made earlier is to say that we're going all in on supporting the full breadth of email. There's a lot of stuff in email that feels ancient that you might, starting out fresh, be like, we shouldn't bother doing this. A good example would be like BCC. Kids these days haven’t heard of a carbon copy, much less a blind carbon copy. It's kind of this weird, esoteric thing. And for a while we didn't have it. We said, we're gonna build a different primitive that's gonna do some of the things that BCC does. And I think what we learned was people are so used to some of these things. And in order to play nicely with existing systems, to play nicely with Gmail, to play nicely with other people's email clients - you really have to support these things fully. You can build cool stuff, but they have to be layered on top. So you can build a nicer interface doing X, Y, Z, but the underlying stuff needs to be like totally standard. And I think we should have accepted that much sooner and said, we are just going to support everything that email does and then build simplifications on top as workflows rather than trying to simplify the underlying model. And three things I learned for the first time: * Gmail pioneered the idea of threaded conversations in email, which was not something email was originally designed for. As a result, Gmail still has a setting where you can disable threads entirely! * Firebase originally started as an app called SendMeHome, which was meant to help people return lost items. The founders pivoted twice, listening to what their users wanted, and eventually landed on Firebase. * HyDE (Hypothetical Document Embeddings) is a RAG technique that involves creating fake documents that might have relevant words that aren't in the document itself (like "leaving" vs "flight" from the quotes above), and using those as a stepping stone to find the right underlying documents. This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.ignorance.ai/subscribe

    41 min

Ratings & Reviews

3
out of 5
2 Ratings

About

Interviews with founders, investors, and deep thinkers on artificial intelligence and its impact on our world. For more deep dives and news stories, visit www.ignorance.ai. www.ignorance.ai