Episode #181: Virginia Eubanks

How does technology, data, and race intersect? On this week’s episode of the show, we talk about how high-tech tools and software profile and punish people of color and low-income people and families. To help me better understand these issues, I speak with Virginia Eubanks about her book, Automating Inequality, and how technology and digital industries perpetuate a permanent underclass.

Virginia Eubanks is an Associate Professor of Political Science at the University at Albany, SUNY. She is the author of Automating Inequality: How High-Tech Tools Profile, Police, and Punish the PoorDigital Dead End: Fighting for Social Justice in the Information Age; and co-editor, with Alethia Jones, of Ain’t Gonna Let Nobody Turn Me Around: Forty Years of Movement Building with Barbara Smith. Her writing about technology and social justice has appeared in Scientific AmericanThe NationHarper’s, and Wired. For two decades, Eubanks has worked in community technology and economic justice movements. She was a founding member of the Our Data Bodies Project and a 2016-2017 Fellow at New America. She lives in Troy, NY.

Episode Notes

Virginia’s website

Virginia’s books:

Cathy O’Neil, Weapons of Math Destruction

Dorothy Roberts at the Harvard Law Review

Our Data Bodies Project

Data4Black Lives

Ella Baker Center

Support the Show

This show is completely listener-supported. There are no ads on the show notes page or in the audio. If you would like to financially support the show, please check out my Patreon page, where just for a few bucks a month, you can get a sneak peek at guests, grab stickers, or even a podcast mug. Your support helps me cover audio editing services, transcription services, and more. You can also support the show by sharing it with others and reviewing it on iTunes or your favorite podcast provider.

Transcript

Welcome back to the PolicyViz podcast. I’m your host Jon Schwabish, and on this week’s episode of the show, we’re going to be talking about how high tech tools and software profile and punish people of color and low income people and families. And to discuss these complex really interesting issues, I chat with Virginia Eubanks who is an associate professor of political science at the University of Albany in New York State. Virginia is also the author of the 2017 book Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor. Now, if you haven’t read Virginia’s book, I really, really highly recommend it. Some of the stories she tells, and she does a really good job of weaving in the data with the stories of people and places, a topic that comes up again in this podcast, is just amazing for me. For me, the book is right up my alley, it’s a mix of public policy considerations, data considerations and technology consideration. So it’s right in that sweet spot. Virginia and I also talk about a lot of groups doing amazing work in this space including the Center for Media Justice, Data 4 Black Lives and the Ella Baker Center, all of which I will link to on the show notes page. And I’m pretty sure you’re going to learn a lot on this week’s episode of the show, so here’s my discussion with Virginia.

Jon Schwabish: Hi Virginia, welcome to the show. Thanks for taking time out of your schedule.

Virginia Eubanks: Yeah, thanks so much for having me.

JS: I’m really excited to chat with you. I really, really enjoyed your book, Automating Inequality. There seem to be a bunch of these books out right now, but the thing that struck me about your book and that I hope we’ll spend some time talking about is you weave in not just the numbers and not just the technical parts of what’s happening but also with the stories and how it impacts real people and real families and real communities. And so, there’s a lot of books out there that I think are a little more academic and your sort of really, for me at least, struck a chord. I’m excited to chat with you about it. I thought maybe you could just talk a little bit about yourself and your background and why you decided to write this particular book on this particular topic.

VE: Yeah. So I’m really glad to hear that the book spoke to you in this way, because it just has always seemed really obvious to me that algorithmic and justice or digital decision making or whatever it is you want to call it right now, particularly in public services, it’s all about people. Right? It’s all about people, and it’s all about politics. And one of the things that I get really concerned about when we talk about these issues sort of in public is that we sometimes frame them, these new technologies, just as issues of sort of administrative upgrades or efficiency upgrades, and so they’re not political in themselves. But one of the arguments I try really hard to make in the book is that these technologies are political decision making machines. And in fact, the thing that is sometimes most worrisome about them is that they’re sort of politics pretending they’re not politics. So this great political scientist that I love, named Deborah Stone, is writing a new book about numbers, and one of the great lines in the book and she says numbers are just stories pretending they’re not stories. And that’s very much sort of the approach I took to this work which is like there’s so much great work out there including sort of earliest, in some ways out, Cathy O’Neil’s wonderful book, Weapons of Mass Destruction. One of the things is so strong about that book is she’s a quant person herself, she’s really good writer, and so she makes it really clear how the technology works and what the impacts might be. But I felt myself after reading it as much as I love the book, really hungry to hear from the people who were being affected, and that goes way, way back in my history. So the moment I think of as the sort of origin story of this book is all the way back in 2000, I was working on a project with a group, a community of women who lived in a residential YWCA in my hometown of Troy, New York, and we were working together around issues of sort of technology and economic inequality. And the sort of idea that was really current at the time was this idea of the digital divide, this idea that the sort of most important social justice issue or one of the most important social justice issues of the digital age was the issue of lack of access, whether that was along racial lines or that was along class lines or gender lines.

And so, I kind of went into this collaborative project in the late 90s with this in my head, this community of wonderful women at the YWCA, really sort of eventually just sat me down and like forced me to have what we in the south would call a come-to-Jesus moment around my assumptions and basically said, look Virginia, we don’t lack interaction with technology in our lives, it’s just the interactions we’re having with them are terrible, are really exploitative, make us feel unsafe, make us feel vulnerable. And one of those moments which I recount very briefly at the beginning of this book feels very much like the seed of Automating Inequality, and that was I was talking to a young mum on public assistance named, goes by a pseudonym in the book Dorothy Allen, and we were talking about her electronic benefits transfer card, her EBT card which is the sort of ATM like card you get public benefits on in most places now. But they were pretty new in 2000, so we were talking about it, and she said I was asking you different questions about how it was working for her, and she said, right, maybe there’s some ways that the stigma is a little bit less than pulling food stamps out in the grocery store, but frankly, most of the tellers don’t know how to use them, so they just shout, like, food stamps, how do I deal with this card. So not that much less stigma. In some ways it’s more convenient, yeah, I guess. But in reality, the thing that really stands out to me, she said, is that when I go to see my caseworker, all of a sudden she’s asking me questions, like, why are you spending all this money at the convenience store on the corner, don’t you know it’s cheaper to go to the grocery store. And so she sort of pointed out that this digital record that was being created by her electronic benefits transfer card was creating a trail that her caseworker could follow to track all of her movements and all of her purchases. And I must have had this incredibly naively shocked look on my face, I was, I don’t know, 25 at the time, and had only been on one public benefit program in the past and not on food stamps. And so, she kind of looked at my face and laughed at me for a while, and then, like God, actually really sort of quiet and concerned, I was like, oh Virginia, you all, meaning, I believe, at the time, meaning sort of professional middle class people, I was a graduate student at the time, like, you all should be paying attention to this because you’re next. And that moment has always stuck in my head not only because I think that was actually incredibly generous of Dorothy, of her being like, oh we’re dealing with this shoot storm, and you all should be concerned because it might impact you too. But also, it stuck in my head this idea that the folks who are sort of on the cutting edge of a lot of the most intrusive invasive digital surveillance technologies are poor working class people, and you need to go to the source to ask people how those tools are operating in their lives.

And so I was really committed writing Automating Inequality, you know, I talked a lot, I did more than a 100 interviews for the book, I talked to lots of different people, I talked to designers, I talked to policymakers, I talked to cops, I talked to frontline social workers, I talked to welfare case examiners, but in every case I started by talking to the people who felt like they were the targets of the system I was describing. So in Indiana, it was folks who either struggled to keep their benefits or lost their benefits during that benefit modernization. In Los Angeles it was unhoused folks who had interacted with the coordinated entry system and either it had gotten them housed and it was often a happy story or they had been shut out somehow. And in Pennsylvania, in Allegheny County story, I started with the families who felt like they were being targeted by this algorithm that risk rates their parenting based on the potential risk to their children of abuse or neglect. And it just turns out that these stories of these magic new digital tools look really different from the point of view of the targets of those tools.

JS: Right.

VE: Yeah, I think it’s just really crucial to start with impact, start with who does it matter to and how’s it affecting their real lives every day.

JS: Yeah, I mean, a lot of what you’re pressing is this idea of empathy through the storytelling, right – being able to put ourselves in someone else’s shoes and say, what if I was the person on snap receiving benefits and having to use this card, what would my experience be like, and maybe that’s something that we’ve lost a little bit over the last, let’s say, three years or so.

VE: Yeah, in a sense, I mean, for me, where that instinct comes from is less about empathy and more about fact. So the old saw is the future has already arrived, it’s just unevenly distributed, this is something that’s widely said, that William Gibson said in the sort of 80s, and I think he meant it slightly differently than I do. I think he meant it that the newest, flashiest technology go to wealthier, more powerful people first. I think in the kinds of cases I’m talking about that these tools are tested first in communities where there’s sort of low expectation that people’s rights will be respected. And so where you see the sort of most bald faced uses of these tools tend to be in these communities, not just poor working class communities and not just communities where people are using public assistance, but I do think that that’s an important place to look, but also migrant communities, communities of color, First Nations, indigenous folks interact with these tools in really different ways than non-native people do. So it’s really about not projecting potential harm of these tools into the future, like, the example would be, and this is actually really important, so let’s talk about, which we do a lot, what an autonomous car would do if it came on in the road upon a box of puppies and a bicyclist, which one would it hit. That’s actually an important question to ask. We should be asking that question. But we have this tendency, when we talk about technology and policy to talk about the problems that might come in the future instead of just going and talking to people about what’s actually happening in their lives right now. And so that just tends to be my approach which is like these future problems are interesting and in some ways sort of beautiful puzzles that people like to sort of grapple with in their heads. But if we want to get real about what’s actually happening we have to go ask people and we have to go ask people in these places where there are real expectations that people’s rights will be respected.

JS: Right. So dealing in the now so that we could deal with the future, we can evolve to the future that we want to get to.

VE: Exactly, yeah, exactly. I think that’s a great way to put it, yeah

JS: So one of the examples, and it’s early in the book is this, you just mentioned this experience in Indiana where the public services mostly TANF, I believe, and food stamps and Medicaid, they were trying to modernize the system, there’s a lot of the technology issues that you talked about throughout the book. I wanted to ask a lot of what you talk about in the book and a lot of what people in the world talk about is a lot of these monetization efforts are about efficiency, they’re about cost cutting, they’re about – and sort of privacy sort of gets a little bit of a wink and a nod of what I see, and I’m just curious how should we as both consumers or people who are receiving these benefits or involved in these programs or just as citizens, so how should we think about these competing incentives because there is a budget constraint for some of these programs, and yet we have these, what I’ll call after reading your book, these various these fairly scary outcomes that are possible?

VE: Yeah. So I think that idea is that we have to work with resources that are limited beyond our ability to change them is one of the most common reasons people will give you for going to these digital tools. So there’s generally two sort of first run reasons that people give for these digital tools. The first is efficiency, cost savings, and sometimes the identification of fraud, waste and abuse. The second is ferreting out legacy patterns of discrimination in frontline decision making. And both of them are reasonable. We want there not to be frontline discrimination in decision making. We want rules to be applied the same way in each case in most cases that we can talk about that some more in a minute because people are individuals and their problems and needs and resources are different. But let’s just talk about the efficiency issue, the triage issue. So even though I spoke to, like I said, 100 people, and I spoke to lots and lots of designers, and all the designers were quite different in their approach, in their politics, and what they thought the problem was – to a person, every single one of them would say that they had to use the tool because it was necessary to do a kind of digital triage, that there weren’t enough resources for everyone, and that in the absence of having sufficient resources they had to make really hard decisions about who should get access to benefits and who shouldn’t. And one of the things that I try to raise in the book is this idea that triage actually isn’t an appropriate way to talk about programs that have been relentlessly defunded. So when you talk about public services, for example, since 1996 and actually even before, we’ve made a series of really consequential political decisions to defund our public service system. And you can’t then say, oh let’s relate this to like a natural disaster. We have to do triage because how could we know that this tsunami was coming, and we don’t have enough medicine. Right? This is clearly not what’s happening. And so one of the arguments I make in the book is that it’s actually not appropriate to use the language of triage, because if the problem is not temporary and if there are not more resources coming, then what you’re doing is not actually triage, it’s digital rationing. And so if it’s digital rationing, let’s name it, let’s say that’s what we’re doing and have a conversation about that. I find this idea that like, oh we have to do it, because we just don’t have enough resources, I find that specious, I find it really unconvincing and potentially mal-intentioned. And you see that across all three of the cases, that’s the same argument that’s made in Indiana and Los Angeles and in Allegheny County. It’s just that there’s just not enough resources. But you end up with these more thorny or more awful problems by defunding these programs at the front end.

So for example, if you look down the road of the book towards the Allegheny County case, so we’re talking now about building a tool that is supposed to risk rate all the families in Allegheny County based on their potential to maltreat that is abuse or neglect their children in the future, so that they can be investigated by the children youth and families, administration of children youth and families there, with an eye for potentially pulling children out of the home and putting them in foster care. The reality is 75% of children who are put in foster care across this country are put in foster care because of neglect, not because of emotional, physical or sexual abuse, and that neglect is basically, the textbook definition of neglect is very similar to just being poor. It means not having safe housing, it means not having enough food, it means having to leave your child alone or with someone who’s not terribly trustworthy because you have to go to work, and all of those are downstream problems from not funding public services. So I find it so sneaky that you then say, oh but we have to do this because we don’t have enough resources to investigate all these dangerous families, when the state is making those families dangerous, it’s not parents who are making the families dangerous, it’s the state. And I mean, even in a very practical way, in Indiana, if this was about cost savings, it did not work because they signed out what was originally a $1.16 billion, 10-year contract, it ended up being a $1.34 billion with a B contract to create a system that basically worked to deny people public assistance, worked so badly that the community rose up and shut it down three years into a 10-year contract. And then IBM turned around and sued the state for breach of contract and originally won, like, won damages on top of the money they had already collected. And if you had just looked at the contract with an eye to how public services actually work and what the impact might be on affected communities, you could tell from the contract out what was going to happen, like, all of the metrics were nothing was like whether or not people got benefits they were entitled to, nothing it was about whether the decision that was made was correct. All the metrics were how fast did you get off the phone, and how many cases did you close. And so you absolutely, I mean, you could have known from the beginning that that was going to be the effect.

JS: Yeah. So that the metrics that they’re looking at are efficiency but not necessarily the usefulness of the program and certainly not to the people who are participating in the programs.

VE: Well, the metrics are short term efficiency, the metrics are like how many people can we get off public assistance this year, and they certainly wouldn’t say that was the metric but I think you could read between the lines of the contract pretty easily that that’s actually the metric. But that just creates, like I said, all of these downstream problems. And I’m not even talking about the human costs, people like Omega Young who lost her Medicaid because she missed a phone appointment because she was in the hospital dying of ovarian cancer. So I’m not even talking about the human effects or the political effects of a community that now will not trust public service programs because they’d have these god awful experiences of being sort of digitally surveilled. I’m not even talking about those, I’m just talking about the straight money, they just lost money on that bet. And I mean, and that’s not even like the legal case that the state had to engage in to fight back against IBM suit that doesn’t include the hundreds and probably thousands of fair hearings that they had to hold when people were wrongly denied their benefits. So not even talking about the cost to people, like the actual cost, they lost money on that.

JS: Right. I’m curious, so the book came out three years ago in 2017, which feels like 90 years ago right now.

VE: It is really funny. It’s really true.

JS: But I’m curious, if at all, how has your perspective changed over the last few months after the murder of George Floyd and Breonna Taylor and unfortunately so many others? How has your perspective changed on these issues especially over the last few months or has it not changed?

VE: Yeah, so I think that you can’t talk about public benefits in the United States without talking about race and without talking about policing and the criminalization of poverty. So I think in this moment I might have framed what I said slightly differently, but I really think so much of the conversations were really exciting and important conversations we’re having about police brutality right now are absolutely clear and obvious in the cases in the book. So though white people still make up the majority of people on public assistance in the United States, perceptions about welfare as a sort of “black thing” impacts all of our policies and all of the ways things are implemented, I mean, everything from racial disparity and foster care to sanction rates in different states, that is sanctioning is throwing people off of public benefits because they’ve made a mistake. All of that is racially determined in some really serious ways. And each case that I talk about, Indiana and Los Angeles and Allegheny County, race plays a really significant role in the case. In Indiana, race played a really significant role on where they rolled out the system first as they were testing it, and actually what I thought was really interesting was that they’re really just a handful of counties in Indiana that have the majority of the Black and African American population, it seems quite intentionally rolled this system out to the counties that did not have Black populations first, which I think is really fascinating. I have some real suspicions that I couldn’t confirm that it might have to deal with sort of using racial resentment as a political tool. So I saw race at play very much in Indiana, and also in the case of folks like Omega Young who really faced the worst outcomes of the system, they were majority Black women. In Los Angeles, I look not just at Skid Row where many of the stories of the unhoused community come out of, and for good reason, it’s a huge and very politically active community; but I also look in South Central which actually has more unhoused people than Skid Row but gets much less funding and much less attention largely because of race. And in Allegheny County, I look at the way that past legacies of racial discrimination are used actually to rationalize the implementation of this tool and sort of the problems with saying that data is racially neutral which I think we’re all pretty familiar with now that we’ve had these sort of conversations about policing, like the way that we stop and frisk, use data, or the way that racism skews data I think is a conversation we’re much more comfortable having these days.

So I mean I feel like that conversation was very much in the book. One of the things that I’m really excited about that has happened since the book is that there have been the sort of intentional linkages that we’ve started to build across different areas of policing, I think of like lowercase policing. So Dorothy Roberts wrote this really great review of the book for the Harvard Law Review that talks about the connections of what I call the digital poorhouse to what she calls the digital carceral state. And it’s exactly what I hoped would happen with the book is that we would start making these connections about how policing operates in different areas, not just in criminal justice and law enforcement, but also in child protection, also in public assistance, also in homeless services, that, as my really brilliant colleague Mariella Saba of Stop LAPD Spying Coalition says, policing wears many uniforms, but that these processes of policing show up in all of these different social programs, and how dangerous that is when you start to conflate economic support programs and law enforcement under the same data structure, under the same rubric and using the same people. I think that’s actually an incredibly dangerous thing.

JS: Yeah. I’m curious, how do you see people, organizations out there trying to work and remedy the problems that you’ve identified – I mean, I feel like we can see police violence, we can see police officers pointing guns at young black men standing at a bus stop which is a story that came out this morning. But the issues that you’re highlighting are some more of these hidden forms of racism and structural racism and have you seen organizations working to turn things around, and how have they sort of gone about doing that?

VE: Yeah, I mean, I think that’s one of the things that makes talking about these technologies so interesting and so important is that we sort of talk about technology particularly as a tool that is neutral that is like you can use it in sort of any old way, I think it’s much more useful to talk about tools as manifestations of structures. So of course the technology that we build for the foster care system is going to be racist, because in every single county in the United States there’s a problem with racial disproportionality in foster care and that’s affected all of the data, and that affects all of the machine learning, and that affects all of the outcomes of all of these tools. And so, it’s like this manifestation of the structural problems that we’re already facing, and I think you’re right though, I think that when we look at these technologies, the harm looks really different than the interpersonal conflict, like, police officer black youth. But the problem that we have is not in the United States is not racist cops, I mean, we do have that problem. The problems though are structural, are really deep, so even if we replaced every police officer tomorrow with Gandhi, we would still have a lot of these problems.

JS: Still have, yeah.

VE: And so that’s the thing about talking about the tech is that it allows us to have those conversations in a way that I think is really, really powerful. So yeah, there is some work going on. I’m probably out of the loop of the sort of newest most exciting work around that. I was involved after the, as the book was coming out and after the book with a really great project called Our Data Bodies project that is starting to sort of imagine what community safety looks like in a world of sort of digital surveillance, the Center for Media Justice – I think they changed their name recently but I don’t remember what it is, which is terrible, sorry guys, you’re awesome – have been doing that work. I think it’s not the kind of work that has data for Black Lives. But it’s not the kind of work that necessarily needs a whole slate of new organizations. It feels to me like it is a layer that we add to the organizing work we’re already doing. So if you’re interested in economic justice, you also have to think about algorithms now. If you’re interested in police brutality, you also have to think about CompStat. If you’re into electronic shackling, that it’s just it’s a dimension of the work that so many people are already taking on. And I think one of the most important things that I really hope people take out of the book is that none of these systems are inevitable. So Indiana is the perfect example. The state was like, you know what, we’re doing this, we don’t care what people say, we’re going to hold basically no public comment, period, on this like more than billion dollar contract, we’re just going to do it, and the citizens of Indiana shut that thing down – they just, old-school organizing, had townhall meetings, went door to door, handed out flyers, and they were just like, no, you don’t get to treat us like this. I think we’ve seen those kinds of successes around facial recognition technology, around tech workers refusing to work on projects that they find morally reprehensible. So we’re seeing that kind of work pop up all over the place. I’m really excited about that moment, the moment that we’re in. One of the things that I really still, I want to say again, because I so want to see it happen, is like this connection between the policing apparatus of law enforcement and the policing apparatus of the programs that we think of as more charitable or more helpful because we have a tendency to think that these tools are okay as long as they’re like “just helping and not punishing anybody”. But the reality is things like Child Protective Services plays both a helping role and a charitable role and a punishment role. And so, if we don’t see that as a policing system, we’re really in danger of giving it a pass around things that we would never accept in law enforcement.

JS: Yeah. There’s this book by Zach Norris who runs the Ella Baker Center out in Oakland called We Keep Us Safe, and he expands on this exact point that a lot of what our public services and programs do is about punishment as opposed to trying to keep people safe and get them, you know, get people who need services to the right place where they can be successful in the long term as opposed to taking kids out of school, putting people behind bars, all sorts of other punishments that we inflict on people in our existing public service programs.

VE: Yeah, well, I love the Ella Baker Center, they’re amazing, and I have heard about this book, I haven’t read it yet, but I’m really excited about it, because one of the things that sort of kept me out of the loop for the last couple of years is, as I started to write the book, my very, very dear and much beloved partner of many years, more than I care to admit, Jason Martin, was attacked and really badly beaten in our neighborhood and ended up suffering from a pretty horrifying case of post-traumatic stress disorder. And living with someone with PTSD, and after the book came out and things calmed down a little bit, we largely sort of turned our attention to his healing and maybe supportive to his healing. And one of the things that’s become really clear to me as a partner of somebody with PTSD and particularly as a partner of somebody with PTSD during a pandemic is how poorly that the systems we hope will work to keep us feeling and actually physically sort of safe and healthy, how routinely and sort of life destroying ways those systems fail us. And I am really interested in figuring out how to live in a world where we keep us safe, and that means something beyond dialing 911, that feels really important to me not just on an intellectual level but on a day to day, you know, leaving my house to go to the corner store level. So it is that kind of safety or security, particularly community security, feels so crucial to me right now, and the pandemic just makes that all the more obvious. The pandemic has made it so clear how harmful these cracks in the system are to everyone, I mean, especially to the people who have already been suffering from falling in the cracks before the pandemic, but the pain has I think been widespread enough through the pandemic that a lot of people are starting to wake up to the cracks in the system.

JS: So I want to ask you one last question on that exact note. So you’ve talked sort of both an optimistic and a pessimistic tone that we are seeing these cracks and there’s a lot of work going on to improve the system and so on. I wonder, going forward, and I know we’ve already talked about, let’s not just go into the future, but looking ahead, are you optimistic that things will change for the better, and probably most of us are feeling sort of down and pessimistic these days, but on the things that we’ve been talking about, are you feeling like things are heading in the right direction or they’re heading in the wrong direction?

VE: So that’s a great question, like, do I think things are getting better or worse, I’m going to say both. And I know what a frustrating answer that is. So I have, for a long time, I have identified myself – and this comes out of the work I did at the YWCA back in the early 2000s – I’ve identified myself as a hard one optimist, in that I feel very aware of the real life shattering system destroying potentially world ending catastrophe we seem to be teetering on the edge of. And also, like, feel like I see so much in the social movements I’ve been part of, and in the community I live in, and in the work that I’ve just been honored to do with all sorts of different folks but primarily with poor and working class communities, I just feel so honored to be in the presence of our continued optimism that we have hope and we can create change, and that out of this mess that we’re in we can birth the kind of world we want to live in. So I don’t think I’m naively optimistic, but I also feel like hope is part of revolution, like, you have to live – Barbara Smith who I did a book with many years ago is a black feminist, who is a great hero and friend of mine, she says, for the revolution to happen you have to live as if the revolution is possible always. And I think, for me, that that feels like a final word.

JS: Yeah. Well, I think that’s a great place to stop.

VE: I always try to give Barbara the last word.

JS: That’s a good quote to end on right there, yeah.

VE: Yeah, that’s right. Well, thanks so much for coming on the show. It was great.

JS: Thank you so much for having me.

Thanks everyone for tuning in to this week’s episode. I hope you learned a lot, I trust you did. If you would like to support the PolicyViz podcast, please consider sharing it on your networks, on Twitter, or leaving a review on your favorite podcast provider platform. Or you can go over to my Patreon page and support the show financially. For just a few bucks a month you get a coffee mug or you get my thanks or you get a note from me every month giving you the heads up on what’s coming up on the podcast. So until next time this has been the PolicyViz podcast. Thanks so much for listening.

A number of people help bring you the PolicyViz podcast. Music is provided by the NRIS. Audio editing is provided by Ken Skaggs. And each episode is transcribed by Jenny Transcription Services. The PolicyViz website is hosted by WP Engine and is published on WordPress. If you’d like to help support the podcast, please visit our Patreon page.

Credits

Music by The NRIs

Audio editing by Ken Skaggs

Transcription by Jenny Transcription Services

PolicyViz.com is hosted by WPEngine and run on WordPress

Photo by Taylor Vick on Unsplash