
In December, the European Union , the first major law aiming to regulate technologies that fall under the umbrella of artificial intelligence. The legislation might have arrived sooner, but the sudden success of ChatGPT in late 2022 demanded the act be updated.
The EU鈥檚 act, however, does not mention fairness 鈥 a measure looking at how well a system avoids discrimination. The field studying fairness in machine learning (a sub-field of AI) is relatively new, so clear regulation is still in development.
Mike Teodorescu, a University of Washington assistant professor in the Information School, proposes in a new paper that private enterprise standards for fairer machine learning systems would inform governmental regulation.
The Feb. 15 by the Brookings Institute as part of its series 鈥.鈥
UW News spoke with Teodorescu about the paper and the field of machine learning fairness.
To start, could you explain what machine learning fairness is?
Mike Teodorescu: It is essentially concerned with ensuring that a machine learning algorithm is fair to all categories of users. It combines computer science, law, philosophy, 快播成人 and some economics as well.
For example, if you鈥檙e trying to create software to automate hiring interviews, you might have a group of HR people interview many candidates with diverse backgrounds and experiences and recommend a binary outcome 鈥 hire or don鈥檛 hire. Data from actual HR interviews can be used to train and test a machine learning model. At the end of this process, you get accuracy 鈥 the percent the model got correct. But this percentage does not capture how well the algorithm performs when considering certain subgroups. U.S. law forbids discrimination based on protected attributes, which include gender, race, age, veteran status and so on.
In the simplest terms, as an example, if you count the number of veterans that you would like to hire, then the algorithm should hire independent of the protected attribute. Of course, this becomes more complex as you have more intersections of subgroups 鈥 you might have race, age, socioeconomic status and gender. From a practical perspective, if you have a system of equalities for dozens of values of protected attributes, it is unlikely that all of them will be satisfied at the same time. I don鈥檛 think we have a generalizable solution and we do not have yet an optimal way to check for AI fairness.
What is it important for the general public to understand about machine learning fairness?
MT: It helps to understand procedural fairness, which looks at the methods that are used to come up with decisions. A user might want to ask, 鈥淒o I know if this software is using machine learning to make some prediction about me? If yes, what kind of inputs is it taking? Can I correct an incorrect prediction? Is there a feedback mechanism by which I can challenge it?鈥
This principle is actually found in privacy laws in Europe and California, where we can object to certain information being used. That level of transparency would be great in the case of a machine learning algorithm being applied to make some decision about you. Maybe there is an option to select what variables it鈥檚 using to show you certain ads. Now, I鈥檓 not sure that鈥檚 something we will see in the very near future, but it鈥檚 something users might care about.
What鈥檚 impeding fairness standards from being widely adopted by companies?
MT: I think it鈥檚 a problem of incentives. From an economic perspective, companies want to bring products to market as quickly as possible. If users get an app that uses image recognition AI, they likely won鈥檛 read the Terms of Service. So they鈥檙e probably not going to spend the time to go through training on whether the tool is fair or not. Many users might not even know that it鈥檚 possible for a tool to be unfair.
For a company right now, the incentive to develop such systems would be to put the company at the technological forefront and to signal quality 鈥 that its AI tools are fairer than its competitors鈥. But if the users do not know about this being a problem, they may not be worried about which company鈥檚 product is fairer. Probably 10 years from now, many more people will care about fairness, just like they do about cybersecurity and data privacy. Cybersecurity wasn鈥檛 such a common concern until we had a lot of these breaches.
Would an example of what you鈥檙e explaining here be somebody submitting a job application to a company that uses a machine learning algorithm to sort applications? That person wouldn鈥檛 necessarily know if there鈥檚 a machine learning algorithm sorting these applications, so they certainly wouldn鈥檛 know if they鈥檝e been unfairly sifted out.
MT: Precisely, and that concern keeps me up at night. There鈥檚 a patchwork of regulations across different countries and states, but there isn鈥檛 yet a comprehensive federal regulation about this. There鈥檚 a law specifically about . There鈥檚 also an EU law that very recently got through, which allows people to contest or determine how their data is being used. There鈥檚 a White House set of directives that have been proposed. Eventually, I think there will be a federal law.
Do you see standards arriving first and then driving actual regulation of machine learning fairness?
MT: Yes, regulations are slow. There are a lot of hurdles to pass a law. But standards play more into the economic incentives. There are standards for cybersecurity, quality measurement, WiFi, Bluetooth and so on, but we don鈥檛 yet quite have accepted standards for machine learning fairness yet. Usually, an organization produces them. The Institute of Electrical and Electronics Engineers (IEEE) comes up with a lot of technical standards, and actually suggested a few. The standards committees within such organizations usually bring people from industry, academia and government together, and they come up with guidelines that can be updated, so there might be different versions of a standard. That provides a lot more flexibility than regulations. For instance, there are two different quality management manufacturing standards. Most factories have the less strict standard, while the stricter standard for medical manufacturing is very expensive and much more difficult to get. In fairness, you might see a light standard and a much more comprehensive one.
Likewise, standards organizations can have auditing requirements. Once a company complies with a standard, there鈥檚 a certain frequency of audits to make sure that the standards continue to be upheld. Having something like that for products that use machine learning would be a great way to improve accountability.
, a digital fellow at Stanford University, was a co-author.
For more information, contact Teodorescu at miketeod@uw.edu.