a16z Podcast - a16z Podcast: It's Not What You Say, It's How You Say It -- When Language Meets Big Data

Episode Date: July 16, 2015

When most people think of big data they think of numbers, but it turns out that a lot of big data -- a lot of the output of our work and activity as humans in fact -- is in the form of words. So what ...can we learn when we apply machine learning and natural language processing techniques to text? The findings may surprise you. For example, did you know that you can predict whether a Kickstarter project will be funded or not based on textual elements alone ... before it's even published? Other findings are not so surprising; e.g., hopefully we all know by now that a word like "synergy" can sink a job description! But what words DO appeal in tech job descriptions when you're trying to draw the most qualified, diverse candidates? And speaking of diversity: What's up with those findings about differences in how men and women describe themselves on their resumes -- or are described by others in their performance reviews? On this episode of the a16z Podcast, Textio co-founder and CEO Kieran Snyder (who has a PhD in linguistics and formerly led product and design in roles at Microsoft and Amazon) shares her findings, answers to some of these questions, and other insights based on several studies they've conducted on language, technology, and document bias.

Transcript
Discussion (0)
Starting point is 00:00:00 Hi, everyone. Welcome to the A16Z podcast. I'm Sonal, and I'm here today with Michael, and we are talking to Kieran Snyder, who is a CEO and co-founder of Textio, a company that analyzes job listings to predict how well they're going to perform and can help optimize them to get more qualified diverse candidates. And interestingly, they've been able to figure out, besides what doesn't work very well in job descriptions, words like synergize, they've been able to figure out what does work well. language like in tech people love to talk about hard problems and tough challenges but it's a lot bigger than just about jobs the ability to understand the words we use and how we use them is pretty important because even though we're completely immersed in a world of tech where a lot of the conversation is around big data as numbers a lot of the data that we produce the output of our work is actually taking place in the form of words and those words matter sometimes how you say things is more influential than what you're actually saying, right? And it's counterintuitive to any of us who've built products before because you like to think you're leading with a strong vision. Clearly words matter. And another place that that plays out is with hidden biases that are often revealed in words. For example, Kieran examined a number of resumes to see the differences between how women and men describes themselves as well as in performance reviews to see the ways that women and men were
Starting point is 00:01:24 described differently. The word abrasive, which has been talked about since then, ended up, you know, being used in 17 out of a couple hundred women's reviews and zero times in men's reviews, right? The sort of stereotypical, like aggressive was used in a man's review with an exhortation to be more of it. And in women's reviews is a term of some judgment. Okay, let's get started. Kieran, welcome. So the reason we actually invited to the ACNC podcast today is because you've been writing a lot of interesting work based on the outcomes of your product where you've been analyzing people's use of language in certain contexts as a way to surface insights. And I think that's really fascinating because I think we have a tendency in our world
Starting point is 00:02:07 to focus on big data as if it's just numbers and not other forms of data because you're really describing, I mean, what you describe your work is doing is applying machine learning to text and natural language. So how does that, how did you kind of, how does that work? And then we can talk a little bit more about how you got there. Yeah, so how does it work? I mean, language is just an encoding of concepts, right? And anything that can be encoded can be measured. And so I was sharing the story the other day. We were actually originally started out looking at Kickstarter projects, right? So we started out with this question, could we just look at the text of a Kickstarter project and some of its metadata around the text and predict, you know, before it was
Starting point is 00:02:50 ever published whether it was going to raise money. And we didn't look at the quality of the idea. We didn't look at whether a celebrity endorsed it. Turns out we got over 90% predictive on minute zero of a project as to whether it was going to hit its fundraising goal based solely on things like how long is the text and what kind of fonts are you using and how many headings do you have. So wait a minute, just to clear unpack that a little bit. So before the project even went live on Kickstarter. Just looking at those features of the text, you were able to predict whether it would be successful or not.
Starting point is 00:03:25 Exactly. What were some of the high-level takeaways from that? Yeah, so longer is better where Kickstarter is concerned, kind of counterintuitive. One thing that broke our hearts, because my co-founder, Jensen Harris and I both have some design background, you would think these cleanly designed projects with this beautiful use of single typography would do best,
Starting point is 00:03:47 not so you want to look like a ransom note so you want to mix and match types you want lots and lots of headings oh my god that sounds visually painful images to be front loaded kind of makes sense but a lot of what we found was not intuitive and so it's just demonstrated for us the value of actually measuring because the whole Kickstarter corpus is out there in the world right so you can actually have great training data you can see how well prior projects have performed and we saw hey we're kind of on to something here, just looking at the, very painful as a product person, the quality of your idea doesn't matter,
Starting point is 00:04:24 just looking at the content aspects we could predict. And how do you account then for all the other sort of outside variables, you know, whether it was at the beginning of the Kickstarter kind of like craze, whether it was a certain time of year for that matter? A certain type of product even. Yeah, or geography. How do you know that, in fact, your analysis was correct? I mean, you can look at some.
Starting point is 00:04:47 of those other factors, right? Because you can see when projects are published. Turns out that doesn't make a big difference. You can see the only things that really move the needle in a very short-term way are do you have a celebrity endorsing you because that can get you a lot of social media attention. It doesn't make or break you, but it can help quite a bit. And generally how good you are your social media strategy can tip the balance a bit. But none of those other factors turned out to be as significant as we.
Starting point is 00:05:17 expected. The ability to really zero end via just the text. Did that surprise you? I mean, we started off with a hypothesis that it would be that way and that, you know, sometimes how you say things is more influential than what you're actually saying, right? And it's counterintuitive to any of us who've built products before because you like to think you're leading with a strong vision. We weren't surprised. We were curious as we started to apply the text. to some other verticals, whether it would extend. You know, our first big area has really been in the area of job listings where we've looked to see the first real product application, where we've looked at listings now from
Starting point is 00:05:59 over 10,000 different companies, we've measured who's applied to which listings. And we do see the content matters. We do see some tailoring by geography. Turns out what works in New York is different than what works in San Francisco. We see a lot of tailoring by industry. So what works to hire in tech is very different than what it looks like to hire a clean adjuster or someone in retail, right? So you see some differentiation, but in all cases, depending on how you're slicing and dicing the categories, that text leads. You know, we've
Starting point is 00:06:27 looked at real estate a little bit prior to launching our jobs application, and we've seen the same principles apply. So far you've been talking about the form of the text, like the length and the fonts and the design, but like, were there particular words that popped out as well in terms of what people said on those Kickstarter descriptions or anything like that. I'm bringing this up because there's just this recent anecdote in the news that I read about someone saying that you can predict the success or default of loan applications based on words people use like God or using God a lot will actually mean you'll default, you're more likely to fall on your loan, for example.
Starting point is 00:07:03 By God, I'll pay you every month, I promise. In Kickstarter, we didn't look at that. We started looking at that for real estate listings and then jobs where we've looked at it quite a bit. So we saw when we were prototyping out the real estate stuff that if you say off street parking, that really moves the needle for low-income homes. But for high-income homes, in terms of the number of people who will go to your open house and then the eventual sale price of your home, for higher-priced homes, it's actually a negative because why would you want to highlight that it has off-street parking? It's just sort of an expectation. So we saw, you know,
Starting point is 00:07:40 vocabulary matter quite a bit. In jobs, it matters huge. hugely. We've identified at this point over 25,000 unique phrases that move the needle on how many people will apply for your job, what demographics, how qualified they are. Could you share some of that insight with us? Because, you know, the reason that came across your work is because I read an article about how you analyzed performance appraisals and job descriptions for insights about what moves the needle and the differences and how people communicate. What are some of the things? I mean, just because we have a huge audience list that does job descriptions. That needs to hire some people.
Starting point is 00:08:18 Yeah, so there's sort of a set of language that works really well for everybody. These are not surprising on the face of them, but when you look, you see lots of them. So things like, we'd love to hear from you, be really encouraging and positive in your listing. Using the right balance of talking to the job seeker, so your background is in science, and you really enjoy roller skating in your free time, and talking about, the company. So we stand for this. So terms of the balance between you statements and we statements can matter. You know, language like in tech, people love to talk about hard problems and tough challenges. Curiously, we see patterns change over time. So my favorite example of this is
Starting point is 00:08:59 the phrase big data. So a year and a half ago, if you use the phrase big data in a tech job listing, it was positive. You know, it was seen as compelling and cutting edge. In June of 2015, it's not negative, but it's totally neutral. That's interesting. I wanted to ask because if everybody sort of gloms on to these best practices, how then does the signal versus the noise shift? Exactly. Marketing content, as with any marketing content,
Starting point is 00:09:26 the patterns that work change as they get popular and get adopted. And so one of the reasons we believe software is so interesting as a solution here is that it can kind of keep track at broad scale of what's actually happening right now in the market. So you may have published a job listing that worked really well a year ago and probably how a lot of your listeners write their job listings as they go back to that one and then they try to edit it and tweak it a little bit and fix it. That's exactly what happens. Right. But it actually doesn't necessarily work because the market has changed. And so there's a lot there.
Starting point is 00:10:00 Were you ever, I mean, I'm just curious about this. Were you ever able to find or study associations between people's intent and outcomes and job listings? So for example, one of the things that we've seen happen a lot is that, people only become real about what they actually want out of a job description when they actually put words to paper and words have that power to sort of help discipline what you're looking for. You might not even know what you're looking for until you write it down. Have you ever looked at anything around that or found heard interesting anecdotes around that, given your work? We have seen that listings tend to perform better when they are originally authored.
Starting point is 00:10:35 So you can see some degradation over time when people patch. you know, I take a little bit from this listing and a little bit from this one and I sort of stitch them together. And it's probably because when you're originally authoring it, you bring that coherent point of view. That's really interesting. So a little bit. It's pretty early for us to have seen that. And we also identify phrases that torpedo you're listing. Like, you know, there are, you know, corporate sort of cliches and jargon. So buzzwords, basically. The biggest, you know, one of the very common, we call it a gateway term that kind of torpedoes your listing is the word synergy.
Starting point is 00:11:09 Oh, my God. That should torpedo any piece of content. I don't care what it is. But it's a gateway term because when people include synergy, they're also significantly more likely to include, you know, value add and make it pop. Right. You know, it's kind of silly. But they're all over the place.
Starting point is 00:11:25 And it turns out every candidate of every demographic group hates them. Yeah. And so there's a lot of opportunity to improve in jobs. In the sort of editorial world, we would call that jargon. And it sounds like... We also call it jargon. I think we all call it jargon. Jargon is jargon, but no, totally.
Starting point is 00:11:44 Actually, it's interesting because with words like that, they're obviously in use because they're useful words. And it's kind of sad because, I mean, Synergy at some point, it's probably a useful word. So it's kind of interesting because over time with your corpus of data, you'll be able to sort of map how people's language changes. And when you think of dictionaries as like these static instruments for capturing text these days,
Starting point is 00:12:05 it is kind of fascinating how language is changing in a way that we're able to track differently now thanks to online and software. It changes lexicography. Yeah. As a whole discipline, it changes lexicography for sure. I don't know that you could do it in a static way anymore. Right, I totally agree.
Starting point is 00:12:19 The internet has just exploded that. Right, exactly. Is there, so if big data is kind of neutral now, is there a kind of job type or job description that's the celebrity of the job search world right now? Yeah, what word is sort of popping out that's really moving the needle for you guys or that you've observed?
Starting point is 00:12:36 There are several. most of your listeners are probably intact. It varies a lot by industry. So at scale right now. Outscale is a very popular phrase. That's popular here too. Yeah, well, it is. You don't want to do things in, use methods that are perceived to be manual or perceived
Starting point is 00:12:54 to be limited in some way. So at scale is one that shines. And it started in tech, but it spread to other industries, which is common that we see that. One of my favorite examples, given that we spend a lot of time talking to HR people is, turns out workforce analytics is no longer a good phrase to use. You want to use people analytics. So you can get these highly specific, you know, deep in an industry changes that if you're in the industry and you're on the cutting edge, you probably know. But if you're just a startup trying to hire your first analytics person, you probably have no idea. You don't have a deep
Starting point is 00:13:33 background in the industry, right? Yeah. So you've described it for job listings in real estate. And so this approach, you think, can extend in different directions. You started with Kickstarter. But what is it that it's doing? And how do you, like, it seems a little bit magical. I have to say that. Like, I know that this is a job listing, so therefore it's going to have to do this. But a real estate listing has to do something kind of different. That's a really good question. So, you know, this is, this approach is as powerful as the data set that you have. So if you want to understand a document type, the very first thing you need to do is collect a lot of examples of the document type. And that means you need the documents and you also need some information about their outcomes.
Starting point is 00:14:17 So you are publishing a Kickstarter project. We want to know, did you make money or not? That's a signal for us. You're publishing a job listing. We want to know, did you attract a lot of good people? Did you attract only men? Did you attract no one? So, you know, for each document type that we take on, the first thing we do is we make sure we build out a great training data set.
Starting point is 00:14:38 And then we apply really classical natural language processing techniques. So we look for patterns. And so we say, okay, these are the ones that we're successful. We're successful is defined as, you know, attracted more applicants than 80% of similar listings maybe. And then we start looking for the linguistic patterns and the successes, the ones that aren't as successful. ones that skew in a certain way, demographically, and then we play that back. So it's sort of a key thing for us is that you get that feedback in real time as you're typing. So as you're working on your document before you ever publish it, for you ever pay to publish it somewhere, you can make it good.
Starting point is 00:15:18 And so the training set is the sort of core of all of that, because without that outcomes data, then it's just someone's opinion. I think, could you extend that to say, like, look, I want to write a screenplay for a blockbuster. I mean, does it, could you, I mean, people probably tried this. In fact, a very prominent Bay Area CEO proposed to us a couple months ago that we started applying this to screenplays. To start writing, to actually start producing content or just analyzing them. Sell it to Hollywood. Oh, wow, that's great.
Starting point is 00:15:48 Yeah, so I think anytime you're writing content to sell something, this is really interesting technology, and you could be selling your company. You could be selling yourself. You're a job seeker with a resume that you want to have optimized. you could be selling your product in an e-commerce setup. You could be marketing yourselves. You could be marketing blast emails. Anytime you're writing content to get people to take an action, this is really
Starting point is 00:16:10 useful technology. Let's talk about where this fits and let's actually go, let's purposely use some jargon here. And let's talk about where it fits in the tech trends, like where it fits in that space. So it sounds like you're describing big data techniques applied to natural language or machine learning techniques applied to natural language. But natural language has been around for over three decades, 30 years. I mean, and in the early days, they didn't have this kind of corpus to train the algorithms
Starting point is 00:16:38 on, obviously. So they had to use different kinds of techniques. Like, where does your work fit? And how do you see how it fits in the evolution of natural language? Like, how is it been and where we are, where are we now, kind of? Yeah. I mean, I think in core natural language processing, empirical strategies have always been really important. So when I was a grad student years ago writing a dissertation, collecting data was just a
Starting point is 00:17:01 lot more work, right? So I had to go and record people in the field and I had to transcribe things. It feels like ancient now, actually, but I actually finished my PhD 12 years ago. It wasn't that ancient. The fact that the Internet has codified everything over the last 15 or 20 years, at least in English and most Western languages, means that you have this ready set of, corpora available for you. The tricky part is collecting the text and the outcomes. The outcomes are the part that's hard. Finding the content is easy. So you're describing the difference between just analyzing something and being able to predict something using that text. Exactly. When you analyze something, you can say, oh, cool, this word is really popular now. That's an interesting fact. It might be valuable
Starting point is 00:17:48 to someone to know it, but it's different than saying this word is actually helping your document in some way. What are some other scenarios where you could use sort of this natural language text analysis to make more predict interesting things? Yeah, so people are really starting to think broadly about this. We saw a New York City-based company helping people optimize the sale of their New York City apartments recently using the right phrases. We've seen people do things in healthcare that I think are really interesting. It's not a known vertical to me, but looking at the kind of notes that doctors take about a patient and predicting that patient's likelihood of having a major insurance incident over the next, you know, 12 to 15 months. It's really interesting things
Starting point is 00:18:34 in actuarial science. I think anytime people are producing text, which by the way, in businesses, whatever your business is, text is actually the thing you produce the most of. Right. I believe that. Which any, any industry. And so people produce a lot of text. It's meant to describe often what they think is going to happen. And so, I mean, the field of opportunity is pretty big. The techniques you're describing, is it the same underlying technique applied to all different domains, but do you have to also train each corpus on a different domain? Like, there's special, like there's inside language in each industry. Or are they, are there also universals across all of them? That's a really good question. You don't know until you train is the
Starting point is 00:19:15 short answer to the question. So we have a set of NLP libraries that look for common attributes of text and we always start out any new vertical by turning them on the on the documents and seeing what happened so things like sentence length almost always interesting things like the density of verbs and adjectives almost always interesting document length almost always interesting but the specific phrases that matter what it means to write a job listing is very different than what it means to predict whether a patient is going to become ill right And so the specifics matter. The goals matter.
Starting point is 00:19:53 So if it's a document that's intended for broad consumption, it really probably shouldn't be longer than 600, 700 words. If it's a stock prospectus where you're giving a company some information about how their stocks are likely to perform, it's going to be pages and pages. And so, you know, the specific benchmarks that you're looking for are often very vertical by vertical, but the principles of the kinds of things you look for. are pretty similar. In the past, it seemed like only really big companies could do this because they had like the type of computing hardware and processing power to pull this off. Like, what's changed that a small startup could do this?
Starting point is 00:20:30 AWS is what has changed things, right? I mean, cloud compute at scale and, you know, Google Cloud and Azure. There's a lot of competitors now, but AWS did this for startups, I think. And I say that not because I worked at Amazon before. But it actually is, like for our team to set up the server infrastructure that we need
Starting point is 00:20:49 trivial, you know, so I think that that's a thing. And just the fact that there's so much text data encoded on the internet, Google has democratized a lot of access to data. And so that has helped to. That's great. Did you guys, I have to ask, did you kind of put any Kickstarter projects up there yourself just to give it a whirl? We were asked this a lot during our fundraising. We did look at pitch decks, by the way. One of the things I will come back to your question. One of the things that's been fascinating about having the beta out there in the world is the ways people are using it. So, of course, they're using it for job listings, but people are using it for everything.
Starting point is 00:21:31 Like, just a couple days ago, I had a material science professor write to me saying, I put all my course syllabi through. And I was like, really? Like, how did that work for you? I can't imagine that that was a good result. And he's like, oh, I threw out all of the job parts. I just looked at gender bias. That was a component that I needed for what I was doing. So describe when you say put it through, like, what happens?
Starting point is 00:21:49 I understand, like, in my head, I have this idea that I'm typing along and, you know, suggestions come flying at me, but... That's exactly what happened. So there's a website and you paste or type in your content. And as you're typing, it's getting annotated and marked up for you with patterns, suggestions, things you might want to change, scores. And you can, in the case of the syllabi, right, you can dial it up or down depending on what you want the outcome to be. So in his case, look, I'm sort of tracking for gender bias. he was looking for a specific aspect of what we provide and that of course the product isn't tuned for what he wants but he still found that aspect to be applicable to what he was doing we're seeing people put marketing content through pitch deck content through so to your question about did we initiate any Kickstarter campaigns we didn't because we weren't making it you guys would be genius at it we might be yes we've given a lot of advice to people on kickstarter projects
Starting point is 00:22:43 since then, but we didn't because we were making an enterprise product, right? And if we had followed through on a Kickstarter product and then it got funded, then we'd have to build it. Right. So what did you find? But we helped friends, for sure. That's great. So what did you find out about the pitch decks? Actually, I'm totally intrigued by that, obviously, given who listens to our podcast. I mean, pitch decks, pitch decks are not always highly text oriented, right? So great pitch decks don't include just your text attributes but there are certainly things like length of your deck that matters slide titles end up
Starting point is 00:23:17 mattering quite a bit because people are looking to see a certain style of content and us let's face it we've all seen any kind of meeting where some one person gets hung up on one word and in a headline yeah it always happens we didn't we didn't go deep on pitch decks but we looked at as many as we could find as we were building our own pitch deck in our last round of funding and found some patterns in the synergy line of questioning were there were there words or phases you should never include in your pitch deck you know I don't know I guess there might not even actually be yeah I wonder if there's never I bet I bet there are we didn't identify them synergy is probably so actually let's talk a little bit more about some of and maybe we should
Starting point is 00:24:01 wrap up on this note let's talk a little bit more about some of your findings around gender differences so you said the material science professor tested his own syllabus which again And I'm not sure that made sense, like you said, because there wasn't a reference corpus to, I guess, I guess. There wasn't, but when you have, you know, tens of thousands of phrases that are lighting up, and he's writing for a science, STEM student population, odds are good that there's going to be some lexical overlap. So, you know, you found some things there. So describe some of your findings around job descriptions, because that's given what your product focuses on right now in terms of gender differences and how people, what things you picked up on. Yeah. So prior to us doing this, there was some really strong qualitative research, right,
Starting point is 00:24:47 that National Coalition of Women in Technology, the Clayman Institute here at Stanford, they've done some really interesting qualitative work. But the number of phrases that they identified was on the order of a couple hundred. Avoid rock star, avoid ninja. You know, we want to hire more women in technology. The interesting things for us, first of all, we've talked to a lot of industries outside of tech. And so while in technology, we want to hire more women. When I talk to people who are hiring ICU nurses or elementary school teachers, bias goes the other way. And so it's very important to us that we don't judge. We just forecast and let you make the right choices for your business. Right. Whatever you're optimizing for, given wherever there's an indifference or
Starting point is 00:25:28 imbalance. So I will say we have validated much of the qualitative research, which is good, that there's, you know, some alignment on those points. We have found cases where things are, it's pretty subtle, right? So the difference between fast-paced environment and rapidly moving environment, it's almost head-scratchingly tiny, but statistically, one of them draws many fewer women to apply. And which one is it, by the way? Are you allowed to tell us? Fast-paced. Oh, okay. Interesting. Fast-paced. So you see sometimes these very fine, distinctions between terms that you can only kind of play out statistically. The other thing I would say is that most individual terms aren't that egregious one way or the other. We put a lot of effort
Starting point is 00:26:15 into making something visual so that you could see patterns. So if you have one sort of male bias term or female bias term, you're probably not going to shift your applicant mix that much. But if you have 10, then you're going to see more substantial impact on your applicant set. Interesting. I feel like this kind of validates the natural language approach because in the past, I think in general people tend to put too much stock only on numbers and not on words, like the whole point of what your work is. But the second part of it is that even, you know, I was thinking about when I was back in grad school, there was a lot of debates between qualitative and quantitative data and what was more valuable. And obviously, at the end of the day, they're both valuable. Exactly. But it's
Starting point is 00:26:56 interesting because for the first time, you're really bringing quantitative, quantitative techniques to something that was traditionally in the qualitative domain, just like conversation analysis. Yes, it's true. I mean, so, you know, I've looked at, in some of my prior research, I've looked at some other document types also. As Textio was getting started, the piece that actually really brought us into jobs was some work I did on performance reviews. So I collected hundreds of performance reviews from men and women who work in technology. They were all voluntarily given, which meant they were all good reviews, which I, was betting on that I was going to be comparing strong performers regardless because you don't
Starting point is 00:27:37 give your review unless it's a good one mostly and I found really striking demographic differences and how men and women who were getting good reviews were described in the language that was used wow that's kind of interesting the the word abrasive which has been talked about since then ended up you know, being used in 17 out of a couple hundred women's reviews and zero times in men's reviews, right? The sort of stereotypical, like aggressive was used in a man's review with an exhortation to be more of it, and in women's reviews is a term of some judgment. And so that was really interesting. I looked more recently at resumes. So I collected 1,100 resumes from men and women in technology, about half of each, and found for men and women,
Starting point is 00:28:26 who have very similar backgrounds, very systematic differences, and how they present themselves in a resume, which is really interesting. How did that difference play out? So men's resumes were shorter. They were much deeper into detail about what they actually produced and worked on. That's kind of counterintuitive because you would think that shorter means you'd be less detailed. But you're saying that they were shorter but more detailed about specific things versus other things.
Starting point is 00:28:54 Women's resumes tended to tell. a story. They were written in prose. They didn't use bullets nearly as much. They included executive summaries. They included detailed statements of their personal interests that were twice as long as what men tended to include. So the women's resumes were stronger on narrative, much lighter on detail. The men's resumes were generally stronger on detail and later on narrative. But one of those kinds of resumes gets flagged as positive much more frequently, in tech especially. We look really for what did they deliver, how quickly and tersely can they communicate. And so as we started looking at some of these documents, we realized that there was
Starting point is 00:29:35 just fascinating opportunities on the job listing front because in these other important business documents, we were seeing demographic differences play out. That's fascinating. Were there any other sort of takeaways you have for people who are job seekers out there who want to optimize a resume based on what you discovered? I mean, the length and the narrative is an interesting point. I mean, does it matter, by the way, for a particular industry? I know you said you did. I looked at tech and resumes. I bet some of the same findings apply in other STEM fields, too.
Starting point is 00:30:06 I bet finance you would see some similar patterns. We do see tech and finance pattern together quite a bit in other document types. That's interesting, by the way, that those two domains. It's the quant. Okay. We love numbers, we love data, we love rigor. Those are things that we share. Yeah, I don't know.
Starting point is 00:30:27 I mean, I'm always very reluctant to tell somebody to change the way they tell their story because I think both of those styles are needed for companies. You need people who can tell a customer story and you need people who can track, find detail. And so I guess I would prefer to tell the story the way you actually are and find the company that values it rather than change the way you tell your story. You're right, because that is actually key. That's a really good point. My one question, though, is this whole idea of, like, optimizing language.
Starting point is 00:30:58 How far does that go? Because at some point, I do think that an optimized becomes average or bland or, you know, not to point fingers, but I'm thinking of demand media, for example. Like, collect all this data on what people like or want or, and then spew out on the other side something that nobody likes or nobody wants. I love that question. So it turns out that if everybody stands the same, no one stands out and looks at. good writers continue to find a way to stand out, and then that changes what works. So the beauty
Starting point is 00:31:27 of a learning system, we used the example of big data before, is that if everybody tries to glom on to the same patterns, they're no longer effective. Someone is going to figure out, as with any marketer, someone is going to figure out how to do it better, and they're going to introduce the next pattern for success. So that's how we see it. What are some other ways that people are using what's out there right now that's been surprising to you? So it was initially, surprising to us that people were using our tool for anything other than job listings because we trained on job listings. We're a data quant oriented company. That's the promise we made to say this is going to help you with job listings. As people started trying new things, we realized,
Starting point is 00:32:09 oh, there's nothing doing quite like what we're doing. So people want to test its limits and see what kinds of content they can put through. Some of the crazier things we've seen, So we've seen resumes, which kind of makes sense. We've seen lots and lots of marketing content. We've seen people putting their product descriptions through. A toy company that recently removed gender labels from their children's toys put toy descriptions through to see if they were still flagging any gender language. Just the ninja toys I expect.
Starting point is 00:32:43 Just the ninja toys, right. And, you know, so it's, we're not trained on those documents, but people are continuing. to use the system for it because I think there's a hunger for that kind of experience and there's nothing tailored to what they need. And so for us, it has offered some insight into what people need and what we might do to support that. Will we be able, will we ever be able to use your technology to ask questions? Like say, hey, I want to know X, Y, or Z based on all the things you've trained it on? I don't know if you'll be able to ask questions, but we do think that you'll be able to use it to generate simple content.
Starting point is 00:33:19 So I think there's something fascinating in the, hey, answer these 10 questions about yourself, tell me where you'd like to work, and I will make the resume that is most likely to get you at a good screening interview. And from there, you're on your own. You've got to be good, right? We're not going to lie for you. We're going to tell you our story, but we're going to tell it in the way that is most likely to get you that job at Google that you want.
Starting point is 00:33:40 Or get that loan down the road, if that's what it is. Right. Avoiding the word God. like conference calls for participation coming through and grant proposals and you know again people are trying to write to get a result so we're seeing quite a bit of variation it's not the majority of the usage but it's you know a few percent all right well that was kieran snider of textio and um another episode of the a six and z podcast thank you everyone

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.