They joke, but I've unironically had lecturers warn me about the need to filter AI results intensively in case they come to racist conclusions.
Algorithms is one of the last classes most CS students take at my UNI. So of course the last couple of days is filled with guest speakers. One of the was a jew talking about racist algorithms in software engineering and racist cameras.
I kept asking pointed questions, pointing out his false assumptions, and generally making a joke out of his presentation. I even got the class to laugh a few times. Do note though, the class was almost entirely white men as this was one of the final, and most difficult, of the CS classes. And you knew the two diversity hires and how they got there (not by merit!).
I later learned that the after-class questionnaire surveys asked why that guy was even there in the first place so often that next year he wasn't asked to join the guest speaker list.
I took that my algorithms class in second year, I thought it was really fun, just the right balance of math and creativity. But yeah, someone always manages to wedge politics into it somewhere. Such a shame.
It was my favorite class. Lots of slick algorithms people have come up with.
While they often come to racist conclusions due to biased sample data, it is funny when a Chinese iris recognition system tells Asians to open their eyes for entry.
That is funny. Or when facial detection systems don't recognise any human faces when black people use them. I've seen that example used to argue why tech giants need to hire more blacks.
facial detection systems don't recognise any human faces when black people use them
iPhones have lidar now, so picking out fat nosed niggers should be easy in the future, although why you would want to let a nigger into a secure area is beyond me.
they will probably regret making a fuss when the police start using the same technology to identify masked niggers
(post is archived)