Category Archives: Ai
How AI works is often a mystery that’s a problem – Nature.com
Download this episode of the Nature Podcast
Many AIs are 'black box' in nature, meaning that part of all of the underlying structure is obfuscated, either intentionally to protect proprietary information, due to the sheer complexity of the model, or both. This can be problematic in situations where people are harmed by decisions made by AI but left without recourse to challenge them.
Many researchers in search of solutions have coalesced around a concept called Explainable AI, but this too has its issues. Notably, that there is no real consensus on what it is or how it should be achieved. So how do we deal with these black boxes? In this podcast, we try to find out.
Subscribe to Nature Briefing, an unmissable daily round-up of science news, opinion and analysis free in your inbox every weekday.
Never miss an episode. Subscribe to the Nature Podcast on Apple Podcasts, Google Podcasts, Spotify or your favourite podcast app. An RSS feed for the Nature Podcast is available too.
The rest is here:
How AI works is often a mystery that's a problem - Nature.com
The year of social media soul-searching: Twitter dies, X and Threads are born and AI gets personal – The Associated Press
We lost Twitter and got X. We tried out Bluesky and Mastodon (well, some of us did). We fretted about AI bots and teen mental health. We cocooned in private chats and scrolled endlessly as we did in years past. For social media users, 2023 was a year of beginnings and endings, with some soul-searching in between.
Heres a look back some of the biggest stories in social media in 2023 and what to watch for next year:
A little more than a year ago, Elon Musk walked into Twitter s San Francisco headquarters, fired its CEO and other top executives and began transforming the social media platform into whats now known as X.
Musk revealed the X logo in July. It quickly replaced Twitters name and its whimsical blue bird icon, online and on the companys San Francisco headquarters.
And soon we shall bid adieu to the twitter brand and, gradually, all the birds, Musk posted on the site.
Because of its public nature and because it attracted public figures, journalists and other high-profile users, Twitter always had an outsized influence on popular culture but that influence seems to be waning.
It had a lot of problems even before Musk took it over, but it was beloved brand with a clear role in the social media landscape, said Jasmine Enberg, a social media analyst at Insider Intelligence. There are still moments of Twitter magic on the platform, like when journalists took the platform to post real-time updates about the OpenAI drama, and the smaller communities on the platform remain important to many users. But the Twitter of the past 17 years is largely gone, and Xs reason for existence is murky.
Since Musks takeover, X has been bombarded by allegations of misinformation and racism, endured significant advertising losses and suffered declines in usage. It didnt help when Musk went on an expletive-ridden rant in an on-stage interview about companies that had halted spending on X. Musk asserted that advertisers that pulled out were engaging in blackmail and, using a profanity, essentially told them to get lost.
Continuing the trend of welcoming back users who had been banned by the former Twitter for hate speech or spreading misinformation, in December, Musk restored the X account of conspiracy theorist Alex Jones, pointing to an unscientific poll he posted to his followers that came out in favor of the Infowars host who repeatedly called the 2012 Sandy Hook school shooting a hoax.
LGBTQ and other organizations supporting marginalized groups, meanwhile, have been raising alarms about X becoming less safe. In April, for instance, it quietly removed a policy against the targeted misgendering or deadnaming of transgender individuals. In June, the advocacy group GLAAD called it the most dangerous platform for LGBTQ people.
GLSEN, an LGBTQ education group, announced in December that it was leaving X, joining other groups such as the suicide prevention nonprofit Trevor Project, saying that Musks changes have birthed a new platform that enables its users to harass and target the LGBTQ+ community without restriction or discipline.
Musks ambitions for X include transforming the platform into an everything app like Chinas WeChat, for instance. The problem? Its not clear if U.S. and Western audiences are keen on the idea. And Musk himself has been pretty vague on the specifics.
While X contends with an identity crisis, some users began looking for a replacement. Mastodon was one contender, along with Bluesky, which actually grew out of Twitter a pet project of former CEO Jack Dorsey, who still sits on its board of directors.
When tens of thousands of people, many of them fed-up Twitter users, began signing up for the (still) invite-only Bluesky in the spring, the app had less than 10 people working on it, said CEO Jay Graber recently.
This meant scrambling to keep everything working, keeping people online, scrambling to add features that we had on the roadmap, she said. For weeks, the work was simply scaling ensuring that the systems could handle the influx.
We had one person on the app for a while, which was very funny, and there were memes about Paul versus all of Twitters engineers, she recalled. I dont think we hired a second app developer until after the crazy growth spurt.
Seeing an opportunity to lure in disgruntled Twitter users, Facebook parent Meta launched its own rival, Threads, in July. It soared to popularity as tens of millions began signing up though keeping people on has been a bit of a challenge. Then, in December, Meta CEO Mark Zuckerberg announced in a surprise move that the company was testing interoperability the idea championed by Mastodon, Bluesky and other decentralized social networks that people should be able to use their accounts on different platforms kind of like your email address or phone number.
Starting a test where posts from Threads accounts will be available on Mastodon and other services that use the ActivityPub protocol, Zuckerberg posted on Threads in December. Making Threads interoperable will give people more choice over how they interact and it will help content reach more people. Im pretty optimistic about this.
Social medias impact on childrens mental health hurtled toward a reckoning this year, with the U.S. surgeon general warning in May that there is not enough evidence to show that social media is safe for children and teens and calling on tech companies, parents and caregivers to take immediate action to protect kids now.
Were asking parents to manage a technology thats rapidly evolving that fundamentally changes how their kids think about themselves, how they build friendships, how they experience the world and technology, by the way, that prior generations never had to manage, Dr. Vivek Murthy told The Associated Press. And were putting all of that on the shoulders of parents, which is just simply not fair.
In October, dozens of U.S. states sued Meta for harming young people and contributing to the youth mental health crisis by knowingly and deliberately designing features on Instagram and Facebook that addict children to its platforms.
In November, Arturo Bjar, a former engineering director at Meta, testified before a Senate subcommittee about social media and the teen mental health crisis, hoping to shed light on how Meta executives, including Zuckerberg, knew about the harms Instagram was causing but chose not to make meaningful changes to address them.
The testimony came amid a bipartisan push in Congress to adopt regulations aimed at protecting children online. In December, the Federal Trade Commission proposed sweeping changes to a decades-old law that regulates how online companies can track and advertise to children, including turning off targeted ads to kids under 13 by default and limiting push notifications.
Your AI friends have arrived but chatbots are just the beginning. Standing in a courtyard at his companys Menlo Park, California headquarters, Zuckerberg said this fall that Meta is focused on building the future of human connection and painted a near-future where people interact with hologram versions of their friends or coworkers and with AI bots built to assist them. The company unveiled an army of AI bots with celebrities such as Snoop Dogg and Paris Hilton lending their faces to play them that social media users can interact with.
Next year, AI will be integrated into virtually every corner of the platforms, Enberg said.
Social apps will use AI to drive usage, ad performance and revenues, subscription sign ups, and commerce activity. AI will deepen both users and advertisers reliance and relationship with social media, but its implementation wont be entirely smooth sailing as consumer and regulatory scrutiny will intensify, she added.
The analyst also sees subscriptions as an increasingly attractive revenue stream for some platforms. Inspired by Musks X, subscriptions started as a way to diversify or boost revenues as social ad businesses took a hit, but they have persisted and expanded even as the social ad market has steadied itself.
With major elections coming up in the U.S. and India among other countries, AIs and social medias role in misinformation will continue to be front and center for social media watchers.
Were not prepared for this, A.J. Nash, vice president of intelligence at the cybersecurity firm ZeroFox, told the AP in May. To me, the big leap forward is the audio and video capabilities that have emerged. When you can do that on a large scale, and distribute it on social platforms, well, its going to have a major impact.
Excerpt from:
Why Bill Gates Says AI Will Supercharge Medical Innovations – CNET
Bill Gates' first grandchild was born in 2023, so the year will forever be special to him, he says. It was also the year that artificial intelligence went mainstream, spurred by the arrival of ChatGPT. And that got Gates thinking about how the world his granddaughter is coming into will change in a positive way because of AI.
The co-founder of Microsoft and a guiding figure of the PC era across several decades, Gates knows a thing or two about technological revolutions. He sees 2024 as a monumental year for artificial intelligence, with the technology becoming especially important in global health, where Gates and his namesake foundation have been working for decades.
"We now have a better sense of what types of jobs AI will be able to do by itself and which ones it will serve as a copilot for," Gates wrote in a lengthy post on his GatesNote blog this week. "And it's clearer than ever how AI can be used to improve access to education, mental health, and more. It motivates me to make sure this technology helps reduce -- and doesn't contribute to -- the awful inequities we see around the world."
It's been quite a year for AI, and more specifically, generative AI. Gen AI goes a step further than other AI methods. It can create new materials, such as text, images, speech or video, based on its own understanding of the patterns it recognizes in data.
Gen AI became known thanks to the launch of OpenAI's ChatGPT in late 2022, although smart home controls and AI-powered virtual assistants such as Alexa had already made inroads into homes and popular culture.
ChatGPT, the frontrunner in the onslaught of generative AI tools released over the last year, allows anyone with a smartphone or a laptop to use AI for generating information or images. These tools been trained on huge swaths of data that allow it to come up with original responses to our queries -- with varying degrees of success. More than 100 million people use ChatGPT each week, OpenAI chief executive Sam Altman said in November. Microsoft is a significant investor in OpenAI.
Other companies aren't ceding territory to Microsoft. In early December, for instance, Google began updating its Bard AI chatbot with a new AI model called Gemini that provides improved text-based chat abilities. Tech companies are continuing to add Gen AI abilities into programs and devices of all kinds, from search engines to smart phones.
In 2023, investors pourednearly $10 billioninto gen AI startups, more than double the $4.4 billion invested the year before, according toGlobalData.
But even as Gen AI explodes in popularity, many users are still cautious. In addition to concerns that AI could replace human employees, many worry about it putting forth inaccurate information. Dictionary.com selected the AI term "hallucinate," describing what happens when AI produces false information, as its word of the year.
Gates thinks mainstream integration of AI is coming soon.
"If I had to make a prediction, in high-income countries like the United States, I would guess that we are 18-24 months away from significant levels of AI use by the general population," he wrote.
But he also sees 2024 as a turning point.
Since stepping down as Microsoft CEO in 2000, Gates has focused on philanthropy, founding the Bill & Melinda Gates Foundation with his now-former wife. It's in areas related to his foundation's work in global health where Gates sees AI becoming helpful in 2024.
Fighting antibiotic resistance: He cites an AI-powered tool under development at the Aurum Institute in Ghana that helps health workers prescribe antibiotics without contributing to antimicrobial resistance, where pathogens learn how to get past antibiotic defenses. The tool can comb through all the available information about antimicrobial resistance and suggest the best drug plan for a patient.
High-risk pregnancy help: A woman dies in childbirth every two minutes, Gates says. He's hopeful that AI can combat this horrifying statistic. AI-powered ultrasounds can help identify pregnancy risks, and the Gates foundation is working to fund that process. Also, AI researchers at ARMMAN, an India-based nonprofit organization, are working on a large language model the technology that underlies ChatGPT and other AI chatbots that can help health workers treating high-risk pregnancies.
HIV risk assessment: Many people aren't comfortable talking to a doctor about their sexual history, but that can be vital for assessing risk for diseases like HIV. Gates is excited about a South African chatbot called Your Choice, being developed by Sophie Pascoe of Wits Health Consortium. The chatbot acts as a nonjudgmental counselor that can provide round-the-clock advice, especially for vulnerable populations.
Quick access to medical records: While people in rich countries may have their medical records easily available, in other countries, many people don't have a documented medical history, Gates says. This can hinder their medical treatment because their doctors need to know about allergies, past health issues and more. A Pakistani team is working on a voice-enabled mobile app that could make this easier, asking a series of questions and filling out a patient's medical record with the answers.
Gates also sees AI assisting in education, calling AI education tools "mindblowing," as they are tailored to individual learners, and says they will "only get better." He's excited about how the technology can be localized to students in many different countries and cultural contexts.
Not everything on Gates' mind is AI-related. He's concerned about climate change, saying he's "blown away by the passion from young climate activists," and hopeful that 2024 will see more investment in innovations that will help those who are most affected by the climate crisis.
And he even plunges into the debate over nuclear energy. Gates notes that high-profile disasters such as Chernobyl in the 1980s and Three Mile Island in the late 1970s have spotlighted the risks, but over the past year, he's seen a shift towards acceptance. He sees the once-bogeyman of the energy world as necessary to meet the world's growing need for energy while eliminating carbon emissions.
A New York Times in early December noted that Gates was "long skeptical" of what AI could do. That changed in August 2022, when he saw a demonstration of OpenAI's GPT-4, the large language model underlying ChatGPT. That sold Gates on the concept, and he helped Microsoft "move aggressively to capitalize on generative AI."
Although Gates left Microsoft's's board in 2020, he's still an adviser to its CEO Satya Nadella. Microsoft has plunged full-bore into the AI world, The company invested heavily in OpenAI, the creator of ChatGPT, earlier this year. And it's been adding the technology across its online services, including its Bing search engine.
The company also reimagined Windows 11 with the addition of Microsoft Copilot, which puts AI assistance always available on the Windows 11 desktop taskbar. Microsoft vice president Yusuf Mehdi calls it the most significant update to the operating system so far, and it works across multiple apps and mobile phones.
Gates' year-end letter compares the rise of AI to that of the internet, email and search engines, noting that it wasn't long ago when many people were unfamiliar with these things, and now they are part of our daily lives. Gates sees the same kind of sea change coming with AI.
But he admits that it won't be smooth, giving an example from his own life.
"I thought I would use AI tools for the foundation's strategy reviews this year, which require reading hundreds of pages of briefing materials that an AI could accurately summarize for me," Gates says.
But that didn't happen.
"Old habits are hard to break, and I ended up preparing for [the reviews] the same way I always do," he writes.
Editors' note: CNET is using an AI engine to help create some stories. For more, see our AI policy.
Read this article:
Why Bill Gates Says AI Will Supercharge Medical Innovations - CNET
AI Says Painting Attributed to Raphael Includes Contributions from Other Artists – ARTnews
A masterpiece hanging in the Museo Nacional del Prado in Madrid has long sparked debates over whether it was the work of Raphael. But a group of researchers now claims to have finally solved the mystery through the use of an artificial intelligence algorithm.
The Madonna della Rosa (Madonna of the Rose) depicts Mary, Joseph, and the baby Jesus, along with an infant version of John the Baptist. Until the 19th century, the painting was attributed to the Italian Renaissance painter Raffaello Sanzio da Urbino, more often known as Raphael. Then doubts were raised over the Joseph figure looking like an afterthought and whether Raphael had painted the lower section.
The museums website page for the oil painting solely credits it to Raphael.
According to a new research paper published on December 21 in the journal Heritage Science, analysis of the painting using an AI algorithm with an accuracy of 98 percent found that the painting was entirely made by the Italian artist. But it raised questions about whether Raphael indeed painted the face of Joseph in the painting.
The researchers, led by University of Bradford visual computer professor Hussan Ugail, noted that the AI analysis supported earlier work by art historians who had previously questioned the full attribution of this painting to Raphael alone, suggesting that his associate, Giulio Romano, might have had a hand in it.
University of Bradford emeritus professor of molecular spectroscopy Howell Edwards, who co-authored the paper, told the Guardian: The AI program analysis of our work has demonstrated conclusively that whereas the three figures of the Madonna, [Jesus] and St John the Baptist are unequivocally by Raphael, that of St Joseph is not, and has been painted by someone else.
In January, Ugail was part of a team of researchers who used AI-assisted computer-based facial recognition on a painting known as the de Brcy Tondo to also help determine it was a work by Raphael. The research team found that the faces of the Madonna and child in the de Brcy Tondo were identical to ones in the Raphael altarpieceSistine Madonna. Then another study called into question the results of that research, and museum experts raised questions about the methodology.
Ugail told the Guardian that he knows nothing about art, and that the reception to his work from art historians can be frosty. I think there is fear and they also think we are naive, that we dont know what we are doing, he said.
While there is ongoing concern over how the use of AI will eliminate the work of human beings, the research team emphasized in the conclusion of their Heritage Science paper that AI could become a useful resource for art historians and collectors as a supplementary tool for verifying paintings alongside existing methods such as scholarly analysis, spectroscopic imaging, and dating techniques.
As advances continue to be made in machine learning and image processing technologies, this method has the potential to become part of an array of tools for artwork analysis and verification, the paper said. It can operate in conjunction with other methods currently in use, including in-depth scrutiny by art historians and various advanced imaging techniques, thus contributing to a more thorough and dependable framework for artwork authentication and analysis.
Continued here:
AI Says Painting Attributed to Raphael Includes Contributions from Other Artists - ARTnews
New AI model can predict human lifespan, researchers say. They want to make sure it’s used for good – Phys.org
This article has been reviewed according to ScienceX's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:
fact-checked
peer-reviewed publication
trusted source
proofread
close
Researchers have created an artificial intelligence tool that uses sequences of life eventssuch as health history, education, job and incometo predict everything from a person's personality to their mortality.
Built using transformer models, which power large language models (LLMs) like ChatGPT, the new tool, life2vec, is trained on a data set pulled from the entire population of Denmark6 million people. The data set was made available only to the researchers by the Danish government.
The tool the researchers built based on this complex set of data is capable of predicting the future, including the lifespan of individuals, with an accuracy that exceeds state-of-the-art models. But despite its predictive power, the team behind the research says it is best used as the foundation for future work, not an end in and of itself.
"Even though we're using prediction to evaluate how good these models are, the tool shouldn't be used for prediction on real people," says Tina Eliassi-Rad, professor of computer science and the inaugural President Joseph E. Aoun Professor at Northeastern University. "It is a prediction model based on a specific data set of a specific population."
close
Eliassi-Rad brought her AI ethics expertise to the project. "These tools allow you to see into your society in a different way: the policies you have, the rules and regulations you have," she says. "You can think of it as a scan of what is happening on the ground."
By involving social scientists in the process of building this tool, the team hopes it brings a human-centered approach to AI development that doesn't lose sight of the humans amid the massive data set their tool has been trained on.
"This model offers a much more comprehensive reflection of the world as it's lived by human beings than many other models," says Sune Lehmann, author on the paper, which was recently published in Nature Computational Science. A Research Briefing on the topic is presented in the same journal issue.
At the heart of life2vec is the massive data set that the researchers used to train their model. The data is held by Statistics Denmark, the central authority on Danish statistics, and, although tightly regulated, can be accessed by some members of the public, including researchers. The reason it's so tightly controlled is it includes a detailed registry of every Danish citizen.
The many events and elements that make up a life and are spelled out in the data, from health factors and education to income. The researchers used that data to create long patterns of recurring life events to feed into their model, taking the transformer model approach used to train LLMs on language and adapting it for a human life represented as a sequence of events.
"The whole story of a human life, in a way, can also be thought of as a giant long sentence of the many things that can happen to a person," says Lehmann, a professor of networks and complexity science at DTU Compute, Technical University of Denmark and previously a postdoctoral fellow at Northeastern.
The model uses the information it learns from observing millions of life event sequences to build what is called vector representations in embedding spaces, where it starts to categorize and draw connections between life events like income, education or health factors. These embedding spaces serve as a foundation for the predictions the model ends up making.
One of the life events that the researchers predicted was a person's probability of mortality.
"When we visualize the space that the model uses to make predictions, it looks like a long cylinder that takes you from low probability of death to high probability of death," Lehmann says. "Then we can show that in the end where there's high probability of death, a lot of those people actually died, and in the end where there's low probability of dying, the causes of death are something that we couldn't predict, like car accidents."
The paper also illustrates how the model is capable of predicting individual answers to a standard personality questionnaire, specifically when it comes to extroversion.
Eliassi-Rad and Lehmann note that although the model makes highly accurate predictions, they are based on correlations, highly specific cultural and societal contexts and the kinds of biases that exist in every data set.
"This kind of tool is like an observatory of societyand not all societies," Eliassi-Rad says. "This study was done in Denmark, and Denmark has its own culture, its own laws and its own societal rules. Whether this can be done in America is a different story."
Given all those caveats, Eliassi-Rad and Lehmann view their predictive model less like an end product and more like the beginning of a conversation. Lehmann says major tech companies have likely been creating these kinds of predictive algorithms for years in locked rooms. He hopes this work can start to create a more open, public understanding of how these tools work, what they are capable of, and how they should and shouldn't be used.
More information: Germans Savcisens et al, Using sequences of life-events to predict human lives, Nature Computational Science (2023). DOI: 10.1038/s43588-023-00573-5
A transformer method that predicts human lives from sequences of life events, Nature Computational Science (2023). DOI: 10.1038/s43588-023-00586-0
Journal information: Nature Computational Science
Original post:
Chamath Palihapitiya says venture capitalists also face disruption from AIand startup founders stand to benefit – Fortune
Artificial intelligence has been inescapable this year. After OpenAI released ChatGPT some 13 months ago, attention turned to how such tools will disrupt careers and industriesand eager venture capitalists poured billions into AI startups that might do the disrupting.
But VCs themselves could get disrupted, according to billionaire investor Chamath Palihapitiya, a former Facebook executive and the CEO of VC firm Social Capital.
We talk about AI as a big disruptor to the big companies and this and that, but AI may be the biggest disruptor to VC in the end, Palihapitiya said on the All-In Podcast this week.
A world where AI proliferates, he said, is positive for founders, who will be able to own more of their companies rather than give away too much equity to VCs.
In the past, he said, a tech startup with $2 million in seed funding might hire seven people and have enough capital to survive for a year and a half, after which it hopefully gained enough traction so that investors would pony up $10 million or $15 million in Series A funding. The downside, of course, is that in exchange for capital, VCs want equity in the company.
But AI tools give founders more leverage, Palihapitiya said, mentioning GitHub Copilot, which makes creating and fixing code much easier. Startups can now hire programmers, perhaps in other countries with lower pay rates, to use such tools to get more done faster, he noted.
The upshot is that, today, a tech startup with the same amount of seed funding might have a three- or four-person team and survive on that $2 million for four years rather than a year a half. Founders could then end up owning 80% of their company with the potential to exit for $50 million or $100 million, and theyve made more money than in a traditional outcome he said.
Its only a matter of time, Palihapitiya added, until they can put two and two together in an Excel spreadsheet to figure out that owning 50% of a $100 million company is greater than owning 18% of some other company when youre massively diluted, or 8% or whatever.
Jason Calacanis, an angel investor, responded that now, instead of founders of a particular cohort competing on who can raise the most money at the highest valuation, hes seen them shifting to, how do I get to profitability and how do I own as much of my company as possible?
Palihapitiya became the face of the SPAC boom-and-bust a few years back due to his involvement with special purpose acquisition companiesshell corporations listed on a stock exchange that acquire a private company, thereby making it public sans the rigors of the IPO process.
This isnt the first time he has mulled the role of VCs in an AI-altered world.
It seems pretty reasonable and logical, he said last month on the podcast, that AI productivity gains will lead to tens or hundreds of millions of startups made up of only one or two people.
Theres a lot of sort of financial engineering that kind of goes away in that world, he said. I think the job of the venture capitalist changes really profoundly. I think theres a reasonable case to make that it doesnt exist.
Excerpt from:
AI Health Coaches Are Coming Soon to a Device Near You – TIME
Ten years ago, the idea of tracking your footsteps or your heartbeat was weird. Those dedicated to the pursuit of quantified self knowledge proselytized in TED Talks, while journalists attended conferences and reported on the strange new trend. Today, over 40% of households in the U.S. own a wearable device, according to statistics service Statista. It is not uncommon to hear retirees comparing or boasting about their step count for the day. The quantified self is ascendant.
Now, as artificial intelligences relentless advance continues, researchers and technologists are looking for ways to take the next stepto build AI health coaches that sift through health data and tell users how to stay fighting fit.
Theres a lot of evidence to suggest that wearables do offer at least some benefits. A review of scientific studies from 2022 found that, across over 160,000 participants in all the studies included, people who were assigned to wear activity trackers took roughly 1,800 more steps each day, which translated to a weight loss of around two pounds.
Wearables change behavior in a number of waysby prompting users to set goals, allowing them to monitor things they care about, by reminding them when theyre not on track to meet their goalssays Carol Maher, a professor of population and digital health at the University of South Australia and a co-author of the review.
These effects often fade with time, however, says Andrew Beam, an assistant professor in the Department of Epidemiology at the Harvard T.H. Chan School of Public Health, who researches medical artificial intelligence.
Accurately detecting the measures that we care about from signal inputsdetermining step count from an wrist-worn accelerometer, for examplerequires AI, but a banal, unsexy type, says Shwetak Patel, professor in computer science and engineering at the University of Washington and director of health technologies at Google. But, he adds, there is much more it can do already do: AI can stretch the capability of that sensor to do things that we may not have thought were possible. This includes features currently available on popular wearable devices, such as fall detection and blood oxygen detection. Some researchers are trying to use the relatively basic health data provided by wearables to detect disease, including COVID-19, although typically not to the same level of accuracy as devices used in clinical settings.
So far, AI has played a supporting role in the rise of the quantified self. Researchers are hoping to make use of recent advances to put AI on center stage.
Patel recently co-authored a paper in which researchers fed data from wearables into large language models, such as OpenAIs GPT series, and had the models output reasoning about the data that could be useful for clinicians seeking to make mental health diagnoses. For example, if a study participants sleep duration data were erratic, the AI system would point this out and then note that erratic sleep patterns can be an indicator of various issues, including stress, anxiety, or other disorders.
The next generation of AI models can reason, says Patel, and this means they could be used for personalized health coaching. (Other researchers argue its not yet clear whether large language models can reason). It's one thing to say, Your average heart rate is 70 beats per minute, he says. But the thing that we're focusing on is how to interpret that. The kind of modeling work we're doing isthe model now knows what 70 beats per minute means in your context.
The data provided by wearables could also allow AI coaches to understand users health at a much greater level of depth than a human coach could, says Patel. For example, a human coach could ask you how you slept, but wearables could provide detailed, objective sleep data.
Maher has also helped author a review of the research on the effectiveness of AI chatbots on lifestyle behaviors, which found that chatbot health coaches can help people increase the amount of physical activity and sleep they get and improve their diets, although the effect was smaller than is typically found for wearables. These studies were done using fairly rudimentary chatbots (developed years ago, well before, for example, OpenAIs ChatGPT) and Maher expects that more sophisticated AI health coaches would be more effective. She notes, however, that there are still challenges that need solving with large language models like ChatGPTsuch as the models tendency to make up information.
There are reasons to be skeptical about chatbot health coaches, says Beam. First, they suffer from the same drop off in effectiveness over time as wearables. Second, in the realm of health, even human scientists given reams of data about an individual do not yet understand enough to give personalized advice.
Even if the evidence doesnt yet exist to offer precise recommendations to different people based on their health data, an AI health coach could monitor whether a given action seems to be helping and adjust its recommendations accordingly. For example, heart rate data during a suggested workout could be used to inform future exercise recommendations, says Sandeep Waraich, product management lead for wearable devices at Google.
Google has not announced plans to launch an AI health coach, although it does plan to provide AI-powered insights to Fitbit users from early 2024, and in August the New York Times reported that Google DeepMind has been working on an AI life adviser. Apple is also reportedly working on an AI health coach, codenamed Quartz, that it plans to release next year.
Its not just the big tech companies that are trying to take data from wearables and provide continuous, personalized health coaching. Health app Humanity claims to be able to determine a user's biological age to within three years based on movement and heart-rate data. Humanitys algorithm was developed using data from the U.K. biobank, which had 100,000 participants wear a wrist-worn accelerometer for a week. But Michael Geer, co-founder and chief strategy officer at Humanity, is more excited about the possibility for tracking how biological age changes. We're not trying to say you're definitely in the body of a 36-year-old. What we're trying to see is basically over time, did [biological age] generally go up or down, and then that's feeding back to figure out what actions are making you healthier or not, he says.
The problem with tracking measures like Humanitys biological age is that there is still no evidence linking those measures to actual health outcomes, like a reduction in all-cause mortality, says Beam. This is a problem with AIs use in health care more broadly, he says. In general, caution is the right approach here. Even within clinical medicine, there's a huge emerging body of literature on how much these AI algorithms know about medicinewe still don't know how that translates to outcomes. We care about outcomes, we care about improving patient health. And there's just a paucity of evidence for that as of now.
Excerpt from:
AI Health Coaches Are Coming Soon to a Device Near You - TIME
2023: A year of groundbreaking advances in AI and computing – Google Research
Posted by Jeff Dean, Chief Scientist, Google DeepMind & Google Research, Demis Hassabis, CEO, Google DeepMind, and James Manyika, SVP, Google Research, Technology & Society
This has been a year of incredible progress in the field of Artificial Intelligence (AI) research and its practical applications.
As ongoing research pushes AI even farther, we look back to our perspective published in January of this year, titled Why we focus on AI (and to what end), where we noted:
We are committed to leading and setting the standard in developing and shipping useful and beneficial applications, applying ethical principles grounded in human values, and evolving our approaches as we learn from research, experience, users, and the wider community.
We also believe that getting AI right which to us involves innovating and delivering widely accessible benefits to people and society, while mitigating its risks must be a collective effort involving us and others, including researchers, developers, users (individuals, businesses, and other organizations), governments, regulators, and citizens.
We are convinced that the AI-enabled innovations we are focused on developing and delivering boldly and responsibly are useful, compelling, and have the potential to assist and improve lives of people everywhere this is what compels us.
In this Year-in-Review post well go over some of Google Research's and Google DeepMinds efforts putting these paragraphs into practice safely throughout 2023.
This was the year generative AI captured the worlds attention, creating imagery, music, stories, and engaging conversation about everything imaginable, at a level of creativity and a speed almost implausible a few years ago.
In February, we first launched Bard, a tool that you can use to explore creative ideas and explain things simply. It can generate text, translate languages, write different kinds of creative content and more.
In May, we watched the results of months and years of our foundational and applied work announced on stage at Google I/O. Principally, this included PaLM 2, a large language model (LLM) that brought together compute-optimal scaling, an improved dataset mixture, and model architecture to excel at advanced reasoning tasks.
By fine-tuning and instruction-tuning PaLM 2 for different purposes, we were able to integrate it into numerous Google products and features, including:
In June, following last years release of our text-to-image generation model Imagen, we released Imagen Editor, which provides the ability to use region masks and natural language prompts to interactively edit generative images to provide much more precise control over the model output.
Later in the year, we released Imagen 2, which improved outputs via a specialized image aesthetics model based on human preferences for qualities such as good lighting, framing, exposure, and sharpness.
In October, we launched a feature that helps people practice speaking and improve their language skills. The key technology that enabled this functionality was a novel deep learning model developed in collaboration with the Google Translate team, called Deep Aligner. This single new model has led to dramatic improvements in alignment quality across all tested language pairs, reducing average alignment error rate from 25% to 5% compared to alignment approaches based on Hidden Markov models (HMMs).
In November, in partnership with YouTube, we announced Lyria, our most advanced AI music generation model to date. We released two experiments designed to open a new playground for creativity, DreamTrack and music AI tools, in concert with YouTubes Principles for partnering with the music industry on AI technology.
Then in December, we launched Gemini, our most capable and general AI model. Gemini was built to be multimodal from the ground up across text, audio, image and videos. Our initial family of Gemini models comes in three different sizes, Nano, Pro, and Ultra. Nano models are our smallest and most efficient models for powering on-device experiences in products like Pixel. The Pro model is highly-capable and best for scaling across a wide range of tasks. The Ultra model is our largest and most capable model for highly complex tasks.
In a technical report about Gemini models, we showed that Gemini Ultras performance exceeds current state-of-the-art results on 30 of the 32 widely-used academic benchmarks used in LLM research and development. With a score of 90.04%, Gemini Ultra was the first model to outperform human experts on MMLU, and achieved a state-of-the-art score of 59.4% on the new MMMU benchmark.
Building on AlphaCode, the first AI system to perform at the level of the median competitor in competitive programming, we introduced AlphaCode 2 powered by a specialized version of Gemini. When evaluated on the same platform as the original AlphaCode, we found that AlphaCode 2 solved 1.7x more problems, and performed better than 85% of competition participants
At the same time, Bard got its biggest upgrade with its use of the Gemini Pro model, making it far more capable at things like understanding, summarizing, reasoning, coding, and planning. In six out of eight benchmarks, Gemini Pro outperformed GPT-3.5, including in MMLU, one of the key standards for measuring large AI models, and GSM8K, which measures grade school math reasoning. Gemini Ultra will come to Bard early next year through Bard Advanced, a new cutting-edge AI experience.
Gemini Pro is also available on Vertex AI, Google Clouds end-to-end AI platform that empowers developers to build applications that can process information across text, code, images, and video. Gemini Pro was also made available in AI Studio in December.
To best illustrate some of Geminis capabilities, we produced a series of short videos with explanations of how Gemini could:
In addition to our advances in products and technologies, weve also made a number of important advancements in the broader fields of machine learning and AI research.
At the heart of the most advanced ML models is the Transformer model architecture, developed by Google researchers in 2017. Originally developed for language, it has proven useful in domains as varied as computer vision, audio, genomics, protein folding, and more. This year, our work on scaling vision transformers demonstrated state-of-the-art results across a wide variety of vision tasks, and has also been useful in building more capable robots.
Expanding the versatility of models requires the ability to perform higher-level and multi-step reasoning. This year, we approached this target following several research tracks. For example, algorithmic prompting is a new method that teaches language models reasoning by demonstrating a sequence of algorithmic steps, which the model can then apply in new contexts. This approach improves accuracy on one middle-school mathematics benchmark from 25.9% to 61.1%.
In the domain of visual question answering, in a collaboration with UC Berkeley researchers, we showed how we could better answer complex visual questions (Is the carriage to the right of the horse?) by combining a visual model with a language model trained to answer visual questions by synthesizing a program to perform multi-step reasoning.
We are now using a general model that understands many aspects of the software development life cycle to automatically generate code review comments, respond to code review comments, make performance-improving suggestions for pieces of code (by learning from past such changes in other contexts), fix code in response to compilation errors, and more.
In a multi-year research collaboration with the Google Maps team, we were able to scale inverse reinforcement learning and apply it to the world-scale problem of improving route suggestions for over 1 billion users. Our work culminated in a 1624% relative improvement in global route match rate, helping to ensure that routes are better aligned with user preferences.
We also continue to work on techniques to improve the inference performance of machine learning models. In work on computationally-friendly approaches to pruning connections in neural networks, we were able to devise an approximation algorithm to the computationally intractable best-subset selection problem that is able to prune 70% of the edges from an image classification model and still retain almost all of the accuracy of the original.
In work on accelerating on-device diffusion models, we were also able to apply a variety of optimizations to attention mechanisms, convolutional kernels, and fusion of operations to make it practical to run high quality image generation models on-device; for example, enabling a photorealistic and high-resolution image of a cute puppy with surrounding flowers to be generated in just 12 seconds on a smartphone.
Advances in capable language and multimodal models have also benefited our robotics research efforts. We combined separately trained language, vision, and robotic control models into PaLM-E, an embodied multi-modal model for robotics, and Robotic Transformer 2 (RT-2), a novel vision-language-action (VLA) model that learns from both web and robotics data, and translates this knowledge into generalized instructions for robotic control.
Furthermore, we showed how language can also be used to control the gait of quadrupedal robots and explored the use of language to help formulate more explicit reward functions to bridge the gap between human language and robotic actions. Then, in Barkour we benchmarked the agility limits of quadrupedal robots.
Designing efficient, robust, and scalable algorithms remains a high priority. This year, our work included: applied and scalable algorithms, market algorithms, system efficiency and optimization, and privacy.
We introduced AlphaDev, an AI system that uses reinforcement learning to discover enhanced computer science algorithms. AlphaDev uncovered a faster algorithm for sorting, a method for ordering data, which led to improvements in the LLVM libc++ sorting library that were up to 70% faster for shorter sequences and about 1.7% faster for sequences exceeding 250,000 elements.
We developed a novel model to predict the properties of large graphs, enabling estimation of performance for large programs. We released a new dataset, TPUGraphs, to accelerate open research in this area, and showed how we can use modern ML to improve ML efficiency.
We developed a new load balancing algorithm for distributing queries to a server, called Prequal, which minimizes a combination of requests-in-flight and estimates the latency. Deployments across several systems have saved CPU, latency, and RAM significantly. We also designed a new analysis framework for the classical caching problem with capacity reservations.
We improved state-of-the-art in clustering and graph algorithms by developing new techniques for computing minimum-cut, approximating correlation clustering, and massively parallel graph clustering. Additionally, we introduced TeraHAC, a novel hierarchical clustering algorithm for trillion-edge graphs, designed a text clustering algorithm for better scalability while maintaining quality, and designed the most efficient algorithm for approximating the Chamfer Distance, the standard similarity function for multi-embedding models, offering >50 speedups over highly-optimized exact algorithms and scaling to billions of points.
We continued optimizing Googles large embedding models (LEMs), which power many of our core products and recommender systems. Some new techniques include Unified Embedding for battle-tested feature representations in web-scale ML systems and Sequential Attention, which uses attention mechanisms to discover high-quality sparse model architectures during training.
Beyond auto-bidding systems, we also studied auction design in other complex settings, such as buy-many mechanisms, auctions for heterogeneous bidders, contract designs, and innovated robust online bidding algorithms. Motivated by the application of generative AI in collaborative creation (e.g., joint ad for advertisers), we proposed a novel token auction model where LLMs bid for influence in the collaborative AI creation. Finally, we show how to mitigate personalization effects in experimental design, which, for example, may cause recommendations to drift over time.
The Chrome Privacy Sandbox, a multi-year collaboration between Google Research and Chrome, has publicly launched several APIs, including for Protected Audience, Topics, and Attribution Reporting. This is a major step in protecting user privacy while supporting the open and free web ecosystem. These efforts have been facilitated by fundamental research on re-identification risk, private streaming computation, optimization of privacy caps and budgets, hierarchical aggregation, and training models with label privacy.
In the not too distant future, there is a very real possibility that AI applied to scientific problems can accelerate the rate of discovery in certain domains by 10 or 100, or more, and lead to major advances in diverse areas including bioengineering, materials science, weather prediction, climate forecasting, neuroscience, genetic medicine, and healthcare.
In Project Green Light, we partnered with 13 cities around the world to help improve traffic flow at intersections and reduce stop-and-go emissions. Early numbers from these partnerships indicate a potential for up to 30% reduction in stops and up to 10% reduction in emissions.
In our contrails work, we analyzed large-scale weather data, historical satellite images, and past flights. We trained an AI model to predict where contrails form and reroute airplanes accordingly. In partnership with American Airlines and Breakthrough Energy, we used this system to demonstrate contrail reduction by 54%.
We are also developing novel technology-driven approaches to help communities with the effects of climate change. For example, we have expanded our flood forecasting coverage to 80 countries, which directly impacts more than 460 million people. We have initiated a number of research efforts to help mitigate the increasing danger of wildfires, including real-time tracking of wildfire boundaries using satellite imagery, and work that improves emergency evacuation plans for communities at risk to rapidly-spreading wildfires. Our partnership with American Forests puts data from our Tree Canopy project to work in their Tree Equity Score platform, helping communities identify and address unequal access to trees.
Finally, we continued to develop better models for weather prediction at longer time horizons. Improving on MetNet and MetNet-2, in this years work on MetNet-3, we now outperform traditional numerical weather simulations up to twenty-four hours. In the area of medium-term, global weather forecasting, our work on GraphCast showed significantly better prediction accuracy for up to 10 days compared to HRES, the most accurate operational deterministic forecast, produced by the European Centre for Medium-Range Weather Forecasts (ECMWF). In collaboration with ECMWF, we released WeatherBench-2, a benchmark for evaluating the accuracy of weather forecasts in a common framework.
The potential of AI to dramatically improve processes in healthcare is significant. Our initial Med-PaLM model was the first model capable of achieving a passing score on the U.S. medical licensing exam. Our more recent Med-PaLM 2 model improved by a further 19%, achieving an expert-level accuracy of 86.5%. These Med-PaLM models are language-based, enable clinicians to ask questions and have a dialogue about complex medical conditions, and are available to healthcare organizations as part of MedLM through Google Cloud.
In the same way our general language models are evolving to handle multiple modalities, we have recently shown research on a multimodal version of Med-PaLM capable of interpreting medical images, textual data, and other modalities, describing a path for how we can realize the exciting potential of AI models to help advance real-world clinical care.
We have also been working on how best to harness AI models in clinical workflows. We have shown that coupling deep learning with interpretability methods can yield new insights for clinicians. We have also shown that self-supervised learning, with careful consideration of privacy, safety, fairness and ethics, can reduce the amount of de-identified data needed to train clinically relevant medical imaging models by 3100, reducing the barriers to adoption of models in real clinical settings. We also released an open source mobile data collection platform for people with chronic disease to provide tools to the community to build their own studies.
AI systems can also discover completely new signals and biomarkers in existing forms of medical data. In work on novel biomarkers discovered in retinal images, we demonstrated that a number of systemic biomarkers spanning several organ systems (e.g., kidney, blood, liver) can be predicted from external eye photos. In other work, we showed that combining retinal images and genomic information helps identify some underlying factors of aging.
In the genomics space, we worked with 119 scientists across 60 institutions to create a new map of the human genome, or pangenome. This more equitable pangenome better represents the genomic diversity of global populations. Building on our ground-breaking AlphaFold work, our work on AlphaMissense this year provides a catalog of predictions for 89% of all 71 million possible missense variants as either likely pathogenic or likely benign.
We also shared an update on progress towards the next generation of AlphaFold. Our latest model can now generate predictions for nearly all molecules in the Protein Data Bank (PDB), frequently reaching atomic accuracy. This unlocks new understanding and significantly improves accuracy in multiple key biomolecule classes, including ligands (small molecules), proteins, nucleic acids (DNA and RNA), and those containing post-translational modifications (PTMs).
On the neuroscience front, we announced a new collaboration with Harvard, Princeton, the NIH, and others to map an entire mouse brain at synaptic resolution, beginning with a first phase that will focus on the hippocampal formation the area of the brain responsible for memory formation, spatial navigation, and other important functions.
Quantum computers have the potential to solve big, real-world problems across science and industry. But to realize that potential, they must be significantly larger than they are today, and they must reliably perform tasks that cannot be performed on classical computers.
This year, we took an important step towards the development of a large-scale, useful quantum computer. Our breakthrough is the first demonstration of quantum error correction, showing that its possible to reduce errors while also increasing the number of qubits. To enable real-world applications, these qubit building blocks must perform more reliably, lowering the error rate from ~1 in 103 typically seen today, to ~1 in 108.
Generative AI is having a transformative impact in a wide range of fields including healthcare, education, security, energy, transportation, manufacturing, and entertainment. Given these advances, the importance of designing technologies consistent with our AI Principles remains a top priority. We also recently published case studies of emerging practices in society-centered AI. And in our annual AI Principles Progress Update, we offer details on how our Responsible AI research is integrated into products and risk management processes.
Proactive design for Responsible AI begins with identifying and documenting potential harms. For example, we recently introduced a three-layered context-based framework for comprehensively evaluating the social and ethical risks of AI systems. During model design, harms can be mitigated with the use of responsible datasets.
We are partnering with Howard University to build high quality African-American English (AAE) datasets to improve our products and make them work well for more people. Our research on globally inclusive cultural representation and our publication of the Monk Skin Tone scale furthers our commitments to equitable representation of all people. The insights we gain and techniques we develop not only help us improve our own models, they also power large-scale studies of representation in popular media to inform and inspire more inclusive content creation around the world.
With advances in generative image models, fair and inclusive representation of people remains a top priority. In the development pipeline, we are working to amplify underrepresented voices and to better integrate social context knowledge. We proactively address potential harms and bias using classifiers and filters, careful dataset analysis, and in-model mitigations such as fine-tuning, reasoning, few-shot prompting, data augmentation and controlled decoding, and our research showed that generative AI enables higher quality safety classifiers to be developed with far less data. We also released a powerful way to better tune models with less data giving developers more control of responsibility challenges in generative AI.
We have developed new state-of-the-art explainability methods to identify the role of training data on model behaviors. By combining training data attribution methods with agile classifiers, we found that we can identify mislabelled training examples. This makes it possible to reduce the noise in training data, leading to significant improvements in model accuracy.
We initiated several efforts to improve safety and transparency about online content. For example, we introduced SynthID, a tool for watermarking and identifying AI-generated images. SynthID is imperceptible to the human eye, doesn't compromise image quality, and allows the watermark to remain detectable, even after modifications like adding filters, changing colors, and saving with various lossy compression schemes.
We also launched About This Image to help people assess the credibility of images, showing information like an image's history, how it's used on other pages, and available metadata about an image. And we explored safety methods that have been developed in other fields, learning from established situations where there is low-risk tolerance.
Privacy remains an essential aspect of our commitment to Responsible AI. We continued improving our state-of-the-art privacy preserving learning algorithm DP-FTRL, developed the DP-Alternating Minimization algorithm (DP-AM) to enable personalized recommendations with rigorous privacy protection, and defined a new general paradigm to reduce the privacy costs for many aggregation and learning tasks. We also proposed a scheme for auditing differentially private machine learning systems.
On the applications front we demonstrated that DP-SGD offers a practical solution in the large model fine-tuning regime and showed that images generated by DP diffusion models are useful for a range of downstream tasks. We proposed a new algorithm for DP training of large embedding models that provides efficient training on TPUs without compromising accuracy.
We also teamed up with a broad group of academic and industrial researchers to organize the first Machine Unlearning Challenge to address the scenario in which training images are forgotten to protect the privacy or rights of individuals. We shared a mechanism for extractable memorization, and participatory systems that give users more control over their sensitive data.
We continued to expand the worlds largest corpus of atypical speech recordings to >1M utterances in Project Euphonia, which enabled us to train a Universal Speech Model to better recognize atypical speech by 37% on real-world benchmarks.
We also built an audiobook recommendation system for students with reading disabilities such as dyslexia.
Our work in adversarial testing engaged community voices from historically marginalized communities. We partnered with groups such as the Equitable AI Research Round Table (EARR) to ensure we represent the diverse communities who use our models and engage with external users to identify potential harms in generative model outputs.
We established a dedicated Google AI Red Team focused on testing AI models and products for security, privacy, and abuse risks. We showed that attacks such as poisoning or adversarial examples can be applied to production models and surface additional risks such as memorization in both image and text generative models. We also demonstrated that defending against such attacks can be challenging, as merely applying defenses can cause other security and privacy leakages. We also introduced model evaluation for extreme risks, such as offensive cyber capabilities or strong manipulation skills.
As we advance the state-of-the-art in ML and AI, we also want to ensure people can understand and apply AI to specific problems. We released MakerSuite (now Google AI Studio), a web-based tool that enables AI developers to quickly iterate and build lightweight AI-powered apps. To help AI engineers better understand and debug AI, we released LIT 1.0, a state-of-the-art, open-source debugger for machine learning models.
Colab, our tool that helps developers and students access powerful computing resources right in their web browser, reached over 10 million users. Weve just added AI-powered code assistance to all users at no cost making Colab an even more helpful and integrated experience in data and ML workflows.
To ensure AI produces accurate knowledge when put to use, we also recently introduced FunSearch, a new approach that generates verifiably true knowledge in mathematical sciences using evolutionary methods and large language models.
For AI engineers and product designers, were updating the People + AI Guidebook with generative AI best practices, and we continue to design AI Explorables, which includes how and why models sometimes make incorrect predictions confidently.
We continue to advance the fields of AI and computer science by publishing much of our work and participating in and organizing conferences. We have published more than 500 papers so far this year, and have strong presences at conferences like ICML (see the Google Research and Google DeepMind posts), ICLR (Google Research, Google DeepMind), NeurIPS (Google Research, Google DeepMind), ICCV, CVPR, ACL, CHI, and Interspeech. We are also working to support researchers around the world, participating in events like the Deep Learning Indaba, Khipu, supporting PhD Fellowships in Latin America, and more. We also worked with partners from 33 academic labs to pool data from 22 different robot types and create the Open X-Embodiment dataset and RT-X model to better advance responsible AI development.
Google has spearheaded an industry-wide effort to develop AI safety benchmarks under the MLCommons standards organization with participation from several major players in the generative AI space including OpenAI, Anthropic, Microsoft, Meta, Hugging Face, and more. Along with others in the industry we also co-founded the Frontier Model Forum (FMF), which is focused on ensuring safe and responsible development of frontier AI models. With our FMF partners and other philanthropic organizations, we launched a $10 million AI Safety Fund to advance research into the ongoing development of the tools for society to effectively test and evaluate the most capable AI models.
In close partnership with Google.org, we worked with the United Nations to build the UN Data Commons for the Sustainable Development Goals, a tool that tracks metrics across the 17 Sustainable Development Goals, and supported projects from NGOs, academic institutions, and social enterprises on using AI to accelerate progress on the SDGs.
The items highlighted in this post are a small fraction of the research work we have done throughout the last year. Find out more at the Google Research and Google DeepMind blogs, and our list of publications.
As multimodal models become even more capable, they will empower people to make incredible progress in areas from science to education to entirely new areas of knowledge.
Progress continues apace, and as the year advances, and our products and research advance as well, people will find more and interesting creative uses for AI.
Ending this Year-in-Review where we began, as we say in Why We Focus on AI (and to what end):
If pursued boldly and responsibly, we believe that AI can be a foundational technology that transforms the lives of people everywhere this is what excites us!
This Year-in-Review is cross-posted on both the Google Research Blog and the Google DeepMind Blog.
Originally posted here:
2023: A year of groundbreaking advances in AI and computing - Google Research
Research at Microsoft 2023: A year of groundbreaking AI advances and discoveries – Microsoft
In this article
It isnt often that researchers at the cutting edge of technology see something that blows their minds. But thats exactly what happened in 2023, when AI experts began interacting with GPT-4, a large language model (LLM) created by researchers at OpenAI that was trained at unprecedented scale.
I saw some mind-blowing capabilities that I thought I wouldnt see for many years, said Ece Kamar, partner research manager at Microsoft, during a podcast recorded in April.
Throughout the year, rapid advances in AI came to dominate the public conversation (opens in new tab), as technology leaders and eventually the general public voiced a mix of wonder and skepticism after experimenting with GPT-4 and related applications. Could we be seeing sparks of artificial general intelligence (opens in new tab)informally defined as AI systems that demonstrate broad capabilities of intelligence, including reasoning, planning, and the ability to learn from experience (opens in new tab)?
While the answer to that question isnt yet clear, we have certainly entered the era of AI, and its bringing profound changes to the way we work and live. In 2023, AI emerged from the lab and delivered everyday innovations that anyone can use. Millions of people now engage with AI-based services like ChatGPT. Copilots (opens in new tab)AI that helps with complex tasks ranging from search to securityare being woven into business software and services.
Underpinning all of this innovation is years of research, including the work of hundreds of world-class researchers at Microsoft, aided by scientists, engineers, and experts across many related fields. In 2023, AIs transition from research to reality began to accelerate, creating more tangible results than ever before. This post looks back at the progress of the past year, highlighting a sampling of the research and strategies that will support even greater progress in 2024.
AI with positive societal impact is the sum of several integral moving parts, including the AI models, the application of these models, and the infrastructure and standards supporting their development and the development of the larger systems they underpin. Microsoft is redefining the state of the art across these areas with improvements to model efficiency, performance, and capability; the introduction of new frameworks and prompting strategies that increase the usability of models; and best practices that contribute to sustainable and responsible AI.
Microsoft uses AI and other advanced technologies to accelerate and transform scientific discovery, empowering researchers worldwide with leading-edge tools. Across global Microsoft research labs, experts in machine learning, quantum physics, molecular biology, and many other disciplines are tackling pressing challenges in the natural and life sciences.
As AI models grow in capability so, too, do opportunities to empower people to achieve more, as demonstrated by Microsoft work in such domains as health and education this year. The companys commitment to positive human impact requires that AI technology be equitable and accessible.
While AI rightly garners much attention in the current research landscape, researchers at Microsoft are still making plenty of progress across a spectrum of technical focus areas.
Cross-company and cross-disciplinary collaboration has always played an important role in research and even more so as AI continues to rapidly advance. Large models driving the progress are components of larger systems that will deliver the value of AI to people. Developing these systems and the frameworks for determining their roles in peoples lives and society requires the knowledge and experience of those who understand the context in which theyll operatedomain experts, academics, the individuals using these systems, and others.
Throughout the year, Microsoft continued to engage with the broader research community on AI and beyond. The companys sponsorship of and participation in key conferences not only showcased its dedication to the application of AI in diverse technological domains but also underscored its unwavering support for cutting-edge advancements and collaborative community involvement.
Microsoft achieved extraordinary milestones in 2023 and will continue pushing the boundaries of innovation to help shape a future where technology serves humanity in remarkable ways. To stay abreast of the latest updates, subscribe to the Microsoft Research Newsletter (opens in new tab) and the Microsoft Research Podcast (opens in new tab). You can also follow us onFacebook (opens in new tab),Instagram (opens in new tab),LinkedIn (opens in new tab),X (opens in new tab), andYouTube (opens in new tab).
Writers, Editors, and ProducersKristina DodgeKate ForsterJessica GartnerAlyssa HughesGretchen HuizingaBrenda PottsChris StetkiewiczLarry West
Managing EditorAmber Tingle
Project ManagerAmanda Melfi
Microsoft Research Global Design LeadNeeltje Berger
Graphic DesignersAdam BlytheHarley Weber
Microsoft Research Creative Studio LeadMatt Corwine
Originally posted here:
Research at Microsoft 2023: A year of groundbreaking AI advances and discoveries - Microsoft
Robotics company using AI that does engineering work by climbing walls – WJAC Johnstown
Robotics company using AI that does engineering work by climbing walls
by Brock Owens
WJAC - GECKO ROBOTICS
Gecko Robotics, based out of Pittsburgh, said it is using AI to find problems before they happen at nuclear reactors, boilers, pipelines, tanks, and ships.
With robots that can climb walls, Gecko Robotics said its mission is to protect some of the world's most important assets.
Company founder and CEO Jake Loosararian said this dream started as a student at Grove City College, and now he's beginning to partner power plants that are using the software he is calling Cantilever.
What those robots are doing is they're gathering information and data as it relates to the health of the structures, Loosararian said. The software was targeted at specifically using our unique data sets that we've been collecting for the last 11 years since I started Gecko out of my college dorm room.
He said making the AI able to climb walls is inspired by a trip to a powerplant in Oil City when he was in college.
The guy who was gathering information to see if the power plant was going to have a forced outage, Loosararian said, Fell and died doing that inspection.
The robot is remote control operated and according to Loosararian takes about half the time doing the engineering jobs. He said it does require some human help.
Loosararian said he understands the fear some people show toward artificial intelligence progress.
I think it's right to be skeptical of technology, but it's more important to prioritize health, safety, and actually doing the job right, Loosararian said.
At least for now, Cantilever should not fully push out the human jobs according to Loosararian.
Loosararian said, Ones that don't adopt these useful pieces of tech are very much at risk of not just providing solutions to the community that they need to be relied on, but also folks jobs are going to be at risk if you can't figure out better ways to actually operate these facilities.
Load more...
More:
Robotics company using AI that does engineering work by climbing walls - WJAC Johnstown