[MUSIC PLAYING] I'm April Williams, and I am an assistant professor in the Department of Communication Media. I'm also faculty in the Digital Studies Institute, and I am an RCI fellow of Research and Community Impact fellow with the NCID. The RCI fellowship is funding my project on online dating. The project is called Swipe Rights, and I am thinking through advocating for people and the way that we use apps and everyday technologies. AI has a lot of bias, whether that's racial bias or gender bias. So my project is really trying to bring awareness to that bias, especially where dating apps are concerned. The dating apps are only giving us what they think we want, and that's the problem. All the social problems that exist in our society get amplified by algorithms. They're not necessarily created by them. And we have to really stop and think through, do I want to use a race filter on Hinge? Do I want to put something in my bio that is racially coded? Do I want to train the algorithm to look for things that might be racially sensitive? Most of the responsibility for bias that happens in dating platforms falls on those who are making the platforms. There's a huge power imbalance where those folks at the top have most of the power, and us individual users have less of the power. I want to empower users to know what they're looking for, to ask questions about where their data is coming from, what happens to their data when they upload it to the apps, and to really call out app designers and platform designers, to get them to be more accountable to users, to be more trustworthy and to be more transparent. I have a forthcoming book which is on online dating and the algorithms that drive the online dating process. It's called Not My Type-- Automating Sexual Racism in Online Dating. And I think about it both from the perspective of online daters and the things that we do, so our individual swiping behaviors, but then also the algorithms that program and drive our swipe-based experiences. I'm working on a guide to accompany the book so that users can know exactly what language they need to use when they are writing to their representatives and saying, hey, look, actually, big tech needs to be regulated. This is something that I care about. This is something that I would vote for. Right? We vote on policy with our dollars, whether that is an in-app purchase, downloading Tinder, downloading Bumble, or any of the other apps that you might use. Everything that we do is really reinforcing demand. So if we want dating apps to sort of stop and think about the racial bias that's in their algorithms, we have to encourage them by saying, no, this is unacceptable. And the way that we do that is by withholding our money from them. So ultimately, the goal would be that dating platforms, dating companies, would do an internal audit and really take accountability for the algorithms that they're currently using. Hopefully, if there were enough users that said, I'm concerned about this, I'm uncomfortable with this, that would really indicate to these programmers to really say, OK, we need to stop and reevaluate the way that we collect data, the way that we use data, the way that we draw inferences about our user base. So with this grant, I'm really pulling together several of the leading scholars in the field and lots of other practitioners as well to think through what it means for the individual user to say to a company, hey, I'm not comfortable with the way that you're using my data. Hey, we're not going to take this anymore.