Is AI Biased against Some Groups and Spreading Misinformation and Extreme Views? – Boston University

Photo by Lidiia Moor/iStock

Artificial Intelligence

Millions of us have played with artificial intelligence tools like ChatGPT, testing their ability to write essays, create art, make films, and improve internet searches. Some have even explored if they can provide friendship and companionshipperhaps a dash of romance.

But can we, should we, trust AI? Is it safe, or is it perpetuating biases and spreading hate and misinformation?

Those are questions that Boston University computer scientist Mark Crovella will investigate with a new project backed by a first-of-its-kind National Science Foundation (NSF) and Department of Energy program. The National Artificial Intelligence Research Resource (NAIRR) Pilot aims to bring a new level of scrutiny to AIs peril and promise by giving 35 projects, including Crovellas, access to advanced supercomputing resources and data.

A BU College of Arts & Sciences professor of computer science and a Faculty of Computing & Data Sciences professor and chair of academic affairs, Crovella will use the high-powered assist to examine a type of AI known as large language models, or LLMs. His goal is to audit LLMsAI programs trained to study and summarize data, produce text and speech, and make predictionsfor socially undesirable behavior. LLMs help drive everything from ChatGPT to automated chatbots to your smart speaker assistant. Crovella will be joined on the project by Evimaria Terzi, a CAS professor of computer science.

According to the NSF, the research resource pilot grew out of President Joe Bidens October 2023 executive order calling for a federally coordinated approach to governing the development and use of AI safely and responsibly.

The Brink asked Crovella about the rapid expansion of AI, how its already part of our everyday lives, and how the NAIRR award will help his team figure out if its trustworthy and safe.

Crovella: Large language models are software tools like ChatGPT that are very rapidly becoming widely used. LLMs are quickly finding uses in educationboth students and teachers use LLMs to accelerate their work; in social settingsmany companies now are selling LLMs as online companions or assistants; and in scienceresearchers use LLMs to find and summarize important developments from the flood of research results published every day. Apple, Microsoft, and Meta have all announced integrations of LLMs into their product lines. In fact, ChatGPT had the fastest uptake of any new software, reaching 100 million users in just two monthsmuch faster than did TikTok or Facebook.

Crovella: Given that millions of people will soon be interacting with LLMs on a daily basis, we think its important to ask questions about what social values these models incorporate. Wed like to know whether such models incorporate biases against protected groups, tendencies to propagate extreme or hateful views, or conversational patterns that steer users toward unreliable information.

Given that millions of people will soon be interacting with LLMs on a daily basis, we think its important to ask questions about what social values these models incorporate.

The question is how to assess these tendencies in a system as complex as ChatGPT. In previous research, weve studied simpler systems from the outside. That is, we gave those systems inputs and observed whether their outputs were biased. However, when we look at LLMs, this strategy starts to break down. For example, weve found cases where an LLM will correctly refuse to answer a question on a sensitive topic when the question is posed in English, but simply by asking the same question in a different language (Portuguese), the model provides an answer that it shouldnt.

So, looking at an LLM just from the outside is not reliable. The genesis of this new project is to ask whether we can look inside an LLM, observe the representations of concepts that it is using, observe the logic that it is following, and, from that, detect whether the system is likely to incorporate undesirable social behaviors. In essence, we want to study an LLM the way a neuroscientist studies a brain in an fMRI machine.

I take Senator Markeys views as emphasizing the need for independent research on AI systems. These fantastically capable systems wouldnt exist without Big Techtheres just no way to marshal the resources needed elsewhere. At the same time, we need a lot of eyes on the problem of whether these new systems are serving us well. I want to point out that our work depends on the fact that some large tech companies have actually open-sourced their modelsgiven their code and data away for free. This has been a socially beneficial act on their part and it has been crucial for the kind of work we are doing.

An LLM is large in two ways: it contains a huge amount of knowledge, obtained from vast training data, and it has an enormously complex system for generating output that makes use of billions of parameters. As a result, the internal representations used in LLMs have been referred to as giant and inscrutable. Just processing the huge amount of information inside a modern LLM is an enormous computational task. Our NAIRR grant gives us access to supercomputing facilities at top national laboratories that will enable us to efficiently analyze the internals of modern LLMs.

There are currently over half a million different LLMs available for the public to use. No doubt in the near future, we will each have our own personalized LLM that will know a lot about us and will help us with many tasks on a minute-to-minute basis. How will we know that these giant and inscrutable systems are trustworthy and safe? Our research is intended to provide methods for answering those questions.

See more here:

Is AI Biased against Some Groups and Spreading Misinformation and Extreme Views? - Boston University

Related Posts

Comments are closed.