Page 821«..1020..820821822823..830840..»

AI and machine learning can successfully diagnose polycystic ovary … – National Institutes of Health (.gov)

News Release

Monday, September 18, 2023

NIH study reviews 25 years of data and finds AI/ML can detect common hormone disorder.

Artificial intelligence (AI) and machine learning (ML) can effectively detect and diagnose Polycystic Ovary Syndrome (PCOS), which is the most common hormone disorder among women, typically between ages 15 and 45, according to a new study by the National Institutes of Health. Researchers systematically reviewed published scientific studies that used AI/ML to analyze data to diagnose and classify PCOS and found that AI/ML based programs were able to successfully detect PCOS.

Given the large burden of under- and mis-diagnosed PCOS in the community and its potentially serious outcomes, we wanted to identify the utility of AI/ML in the identification of patients that may be at risk for PCOS, said Janet Hall, M.D., senior investigator and endocrinologist at the National Institute of Environmental Health Sciences (NIEHS), part of NIH, and a study co-author. The effectiveness of AI and machine learning in detecting PCOS was even more impressive than we had thought.

PCOS occurs when the ovaries do not work properly, and in many cases, is accompanied by elevated levels of testosterone. The disorder can cause irregular periods, acne, extra facial hair, or hair loss from the head. Women with PCOS are often at an increased risk for developing type 2 diabetes, as well as sleep, psychological, cardiovascular, and other reproductive disorders such as uterine cancer and infertility.

PCOS can be challenging to diagnose given its overlap with other conditions, said Skand Shekhar, M.D., senior author of the study and assistant research physician and endocrinologist at the NIEHS. These data reflect the untapped potential of incorporating AI/ML in electronic health records and other clinical settings to improve the diagnosis and care of women with PCOS.

Study authors suggested integrating large population-based studies with electronic health datasets and analyzing common laboratory tests to identify sensitive diagnostic biomarkers that can facilitate the diagnosis of PCOS.

Diagnosis is based on widely accepted standardized criteria that have evolved over the years, but typically includes clinical features (e.g., acne, excess hair growth, and irregular periods) accompanied by laboratory (e.g., high blood testosterone) and radiological findings (e.g., multiple small cysts and increased ovarian volume on ovarian ultrasound). However, because some of the features of PCOS can co-occur with other disorders such as obesity, diabetes, and cardiometabolic disorders, it frequently goes unrecognized.

AI refers to the use of computer-based systems or tools to mimic human intelligence and to help make decisions or predictions. ML is a subdivision of AI focused on learning from previous events and applying this knowledge to future decision-making. AI can process massive amounts of distinct data, such as that derived from electronic health records, making it an ideal aid in the diagnosis of difficult to diagnose disorders like PCOS.

The researchers conducted a systematic review of all peer-reviewed studies published on this topic for the past 25 years (1997-2022) that used AI/ML to detect PCOS. With the help of an experienced NIH librarian, the researchers identified potentially eligible studies. In total, they screened 135 studies and included 31 in this paper. All studies were observational and assessed the use of AI/ML technologies on patient diagnosis. Ultrasound images were included in about half the studies. The average age of the participants in the studies was 29.

Among the 10 studies that used standardized diagnostic criteria to diagnose PCOS, the accuracy of detection ranged from 80-90%.

Across a range of diagnostic and classification modalities, there was an extremely high performance of AI/ML in detecting PCOS, which is the most important takeaway of our study, said Shekhar.

The authors note that AI/ML based programs have the potential to significantly enhance our capability to identify women with PCOS early, with associated cost savings and a reduced burden of PCOS on patients and on the health system.

Follow-up studies with robust validation and testing practices will allow for the smooth integration of AI/ML for chronic health conditions.

Several NIEHS clinical studies focus on understanding and detecting PCOS. Learn more athttps://joinastudy.niehs.nih.gov.

Grants. This work was supported by the Intramural Research Program of the NIH/National Institute of Environmental Health Sciences (ZIDES102465 and ZIDES103323).

About the National Institute of Environmental Health Sciences (NIEHS): NIEHS supports research to understand the effects of the environment on human health and is part of the National Institutes of Health. For more information on NIEHS or environmental health topics, visit https://www.niehs.nih.govor subscribe to a news list.

About the National Institutes of Health (NIH):NIH, the nation's medical research agency, includes 27 Institutes and Centers and is a component of the U.S. Department of Health and Human Services. NIH is the primary federal agency conducting and supporting basic, clinical, and translational medical research, and is investigating the causes, treatments, and cures for both common and rare diseases. For more information about NIH and its programs, visit http://www.nih.gov.

NIHTurning Discovery Into Health

Barrera FJ, Brown EDL, Rojo A, Obeso J, Plata H, Lincango EP, Terry N, Rodrguez-Gutirrez R, Hall JE, Shekhar S, 2023. Application of Machine Learning and Artificial Intelligence in the Diagnosis and Classification of Polycystic Ovarian Syndrome: A Systematic Review. Frontiers in Endocrinology.https://www.frontiersin.org/articles/10.3389/fendo.2023.1106625/full

###

See the original post here:
AI and machine learning can successfully diagnose polycystic ovary ... - National Institutes of Health (.gov)

Read More..

Role of Machine Learning in Predicting Disease Outbreaks – Analytics Insight

In this article we try to understand how machine learning can be beneficial for predicting disease outbreaks

By 2050, the 7.8 billion people who currently live on the planet are projected to number 9.7 billion. Unfortunately, the rate of infectious diseases is rising as a result of population rise. The development of diseases is influenced by several variables. These include urbanization, globalization, and climate change, and most of these phenomena are to some degree a result of human activity. Pathogens themselves may be prone to emerging, and the emerging pathogens tend to have more quickly evolving viruses. When a virus from one person infects a different person or an animal, an infectious illness ensues.

It can impact society on a large scale, similar to the coronavirus COVID-19, and is consequently a significant societal issue. It is viewed as a societal issue since it not only negatively affects people but also negatively affects society as a whole. Therefore, identifying high-risk locations for fatal infectious and non-infectious disease outbreaks is crucial in order to undertake prediction and detection of deadly disease outbreaks and improve the effectiveness of the response to these deadly disease outbreaks. To stop the spread of catastrophic infectious disease outbreaks (like COVID-19), health officials can use a variety of machine learning (ML) technologies.

This may be achieved by utilizing machine learning algorithms for both forecasting and identifying lethal infectious diseases as well as for responding to them. In order to predict disease outbreaks, machine learning algorithms can be used to learn datasets that include details about known viruses, animal populations, human demographics, biology and biodiversity data, readily accessible physical infrastructures, cultural/social practices around the world, and the geographic locations of the diseases. For instance, Support Vector Machine (SVM) and Artificial Neural Network (ANN) models can be used to predict malaria outbreaks.

Average monthly precipitation, temperature, humidity, the total number of positive cases, the total number of Plasmodium Falciparum (pF) cases, and the binary values of the number of outbreaks each month Yes or No, as the models performance is evaluated using their predictors, Root Mean Square Error (RMSE), and Receiver Operating Characteristic (ROC). The machine learning techniques may be integrated into an intelligent system to assess or mine social media data for any indicators of any outliers associated to uncommon flu symptoms in order to build efficient diagnostic methods.

As an illustration, Chae et al. suggested using deep learning to forecast infectious illnesses. In their study, deep learning algorithm parameters are improved while simultaneously using social media data to improve detection performance. The metrics include things like the frequency of verified infectious disease diagnoses, the volume of daily Google searches, the volume of Twitter mentions of the illness, and the average temperature and humidity in South Korea as a whole. After a disease event is recognized, responding to infectious disease outbreaks requires making a swift, well-informed choice in order to minimize the harm caused by the effect of the disease outbreaks.

In order to forecast the location and pace of spread of a disease, machine learning techniques may also learn integrated multi-source data from travel itinerary, population, logistics, and epidemiological data. Machine learning techniques can be applied by medical professionals to enhance the delivery of current therapies and hasten the development of novel ones. For instance, to learn any medical data gathered by the hospitals, they may utilize deep learning algorithms to model massive data sets. For instance, data from coronavirus patient clinical testing may be used as input for machine learning models to help clinicians identify the disease more quickly.

View original post here:
Role of Machine Learning in Predicting Disease Outbreaks - Analytics Insight

Read More..

Neurosnap: Revolutionizing Biology Research with Machine Learning – Yahoo Finance

Wilmington, Delaware--(Newsfile Corp. - September 22, 2023) - In a breakthrough development for the field of computational biology, a new startup named Neurosnap is making waves with its innovative approach to incorporating machine learning into the world of biology research. By providing easy access to state-of-the-art bioinformatic tools and models without requiring any coding or technical expertise, Neurosnap aims to accelerate scientific discoveries and advancements in synthetic biology, pharmaceuticals, and medical research.

The marriage of machine learning and biology has shown great promise in recent years, with tools like AlphaFold2 ushering in a new era of possibilities for biologists. However, such cutting-edge tools have often been inaccessible to many researchers, primarily due to the complexity involved in integrating them into their existing pipelines. Neurosnap seeks to address this crucial barrier by offering a fully end-to-end suite of machine learning tools that are user-friendly and seamlessly integrate with a variety of research pipelines.

"Neurosnap was born out of the belief that computational biology has the potential to transform the way we understand and approach complex biological processes," says Keaun Amani, the CEO and co-founder of Neurosnap. "Our mission is to empower researchers from diverse backgrounds to harness the power of machine learning without the burden of technical intricacies. We envision a future where groundbreaking discoveries are made possible by democratizing access to advanced bioinformatic tools."

One of the key features that sets Neurosnap apart is its user-friendly interface, allowing researchers with little to no prior experience in machine learning to leverage its capabilities effectively. By eliminating the need for coding expertise, the platform ensures that biologists can focus on their core research questions and spend less time grappling with the complexities of data analysis.

Story continues

Researchers using Neurosnap can now explore intricate biological phenomena, analyze complex genomic datasets, and predict protein structures with ease. The platform leverages the latest advancements in machine learning algorithms to assist biologists in unraveling the mysteries of life more efficiently than ever before.

The potential impact of Neurosnap on the pharmaceutical and medical fields is particularly promising. By enabling researchers to identify potential drug candidates, predict protein interactions, and analyze disease-related pathways at a faster pace, the platform holds the potential to accelerate drug discovery and development timelines significantly.

With the launch of Neurosnap, the future of computational biology looks brighter than ever. As researchers from diverse backgrounds unite under a common platform, the potential for scientific advancements in various fields of biology becomes limitless. By democratizing access to cutting-edge machine learning tools, Neurosnap is poised to revolutionize the way biological research is conducted.

Name: Keaun AmaniEmail: hello@neurosnap.ai

To view the source version of this press release, please visit https://www.newsfilecorp.com/release/181481

Read more here:
Neurosnap: Revolutionizing Biology Research with Machine Learning - Yahoo Finance

Read More..

Introduction to Deep Learning Libraries: PyTorch and Lightning AI – KDnuggets

Deep learning is a branch of the machine learning model based on neural networks. In the other machine model, the data processing to find the meaningful features is often done manually or relying on domain expertise; however, deep learning can mimic the human brain to discover the essential features, increasing the model performance.

There are many applications for deep learning models, including facial recognition, fraud detection, speech-to-text, text generation, and many more. Deep learning has become a standard approach in many advanced machine learning applications, and we have nothing to lose by learning about them.

To develop this deep learning model, there are various library frameworks we can rely upon rather than working from scratch. In this article, we will discuss two different libraries we can use to develop deep learning models: PyTorch and Lighting AI. Lets get into it.

PyTorch is an open-source library framework to train deep-learning neural networks. PyTorch was developed by the Meta group in 2016 and has grown in popularity. The rise of popularity was thanks to the PyTorch feature that combines the GPU backend library from Torch with Python language. This combination makes the package easy to follow by the user but still powerful in developing the deep learning model.

There are a few standout PyTorch features that are enabled by the libraries, including a nice front-end, distributed training, and a fast and flexible experimentation process. Because there are many PyTorch users, the community development and investment were also massive. That is why learning PyTorch would be beneficial in the long run.

PyTorch building block is a tensor, a multi-dimensional array used to encode all the input, output, and model parameters. You can imagine a tensor like the NumPy array but with the capability to run on GPU.

Lets try out the PyTorch library. Its recommended to perform the tutorial in the cloud, such as Google Colab if you dont have access to a GPU system (although it could still work with a CPU). But, If you want to start in the local, we need to install the library via this page. Select the appropriate system and specification you have.

For example, the code below is for pip installation if you have a CUDA-Capable system.

After the installation finishes, lets try some PyTorch capabilities to develop the deep learning model. We will do a simple image classification model with PyTorch in this tutorial based on their web tutorial. We would walk on the code and have an explanation of what happened within the code.

First, we would download the dataset with PyTorch. For this example, we would use the MNIST dataset, which is the number handwritten classification dataset.

We download both the MNIST train and test datasets to our root folder. Lets see what our dataset looks like.

Every image is a single-digit number between zero and nine, meaning we have ten labels. Next, lets develop an image classifier based on this dataset.

We need to transform the image dataset into a tensor to develop a deep learning model with PyTorch. As our image is a PIL object, we can use the PyTorch ToTensor function to perform the transformation. Additionally, we can automatically transform the image with the datasets function.

By passing the transformation function to the transform parameter, we can control what the data would be like. Next, we would wrap the data into the DataLoader object so the PyTorch model could access our image data.

In the code above, we create a DataLoader object for the train and test data. Each data batch iteration would return 64 features and labels in the object above. Additionally, the shape of our image is 28 * 28 (height * width).

Next, we would develop the Neural Network model object.

In the object above, we create a Neural Model with few layer structure. To develop the Neural Model object, we use the subclassing method with the nn.module function and create the neural network layers within the__init__.

We initially convert the 2D image data into pixel values inside the layer with the flatten function. Then, we use the sequential function to wrap our layer into a sequence of layers. Inside the sequential function, we have our model layer:

By sequence, what happens above is:

Lastly, the forward function is present for the actual input process for the model. Next, the model would need a loss function and optimization function.

For the next code, we just prepare the training and test preparation before we run the modeling activity.

Now we are ready to run our model training. We would decide how many epochs (iterations) we want to perform with our model. For this example, lets say we want it to run for five times.

The model now has finished their training and able to be used for any image prediction activity. The result could vary, so expect different results from the above image.

Its just a few things that PyTorch can do, but you can see that building a model with PyTorch is easy. If you are interested in the pre-trained model, PyTorch has a hub you can access.

Lighting AI is a company that provides various products to minimize the time to train the PyTorch deep learning model and simplify it. One of their open-source product is PyTorch Lighting, which is a library that offers a framework to train and deploy the PyTorch model.

Lighting offers a few features, including code flexibility, no boilerplate, minimal API, and improved team collaboration. Lighting also offers features such as multi-GPU utilization and swift, low-precision training. This made Lighting a good alternative to develop our PyTorch model.

Lets try out the model development with Lighting. To start, we need to install the package.

With the Lighting installed, we would also install another Lighting AI product called TorchMetrics to simplify the metric selection.

With all the libraries installed, we would try to develop the same model from our previous example using a Lighting wrapper. Below is the whole code for developing the model.

Lets break down what happen in the code above. The difference with the PyTorch model we developed previously is that the NNModel class now uses subclassing from the LightingModule. Additionally, we assign the accuracy metrics to assess using the TorchMetrics. Then, we added the training and testing step within the class and set up the optimization function.

With all the models set, we would run the model training using the transformed DataLoader object to train our model.

With the Lighting library, we can easily tweak the structure you need. For further reading, you could read their documentation.

PyTorch is a library for developing deep learning models, and it provides an easy framework for us to access many advanced APIs. Lighting AI also supports the library, which provides a framework to simplify the model development and enhance the development flexibility. This article introduced us to both the library's features and simple code implementation.Cornellius Yudha Wijaya is a data science assistant manager and data writer. While working full-time at Allianz Indonesia, he loves to share Python and Data tips via social media and writing media.

Go here to see the original:
Introduction to Deep Learning Libraries: PyTorch and Lightning AI - KDnuggets

Read More..

Machine learning, but in space | astrobites – Astrobites

This guest post was written by Mingchen (Michelle) Wang, an undergraduate studying Statistics and Data Science at UCLA. She is passionate to explore fields where statistical analysis and astronomy combine. Mingchen wrote this article as part of a science journalism class taught by Astrobites author, Briley Lewis.

Title: A Measurement of the Kuiper Belts Mean Plane From Objects Classified By Machine Learning

Authors: Ian Matheson, Renu Malhotra

First Authors Institution: Department of Aerospace & Mechanical Engineering, University of Arizona, Tucson, AZ 85721, USA

Status: Posted to arXiv

A collection of small icy bodies form a donut-like shape that orbits around the sun. This donut, known as the Kuiper Belt, lies further than the orbit of Neptune in our solar system. Like the Asteroid Belt, but far larger, it is a region of leftovers from the solar systems early history. It is home to millions of Kuiper Belt objects (KBOs), trans-Neptunian objects (TNOs), and also to many dwarf planets, including Pluto, Orcus, Haumea, and Makemake.

For decades, scientists have extensively studied the Kuiper Belt. Previous measurements concluded that this belt stretches from roughly 30 to 55 AU, but that is a lot of uncertainty! For reference, one astronomical unit (AU) is about 93 million miles. An uncertainty of 1860 million miles doesnt seem too good, right? Well, you are in luck! This paper walks through a concept of finding an accurate measurement for the size of the Kuiper Belt, specifically using machine learning techniques. After all, machine learning is great at handling large amounts of data and simulations, just like ChatGPT.

When we have millions of particles in the Kuiper Belt plane, how do we determine where each one of them may be? This is when a bit of data modeling comes into play. Previous researchers focused on a theory that introduced the idea of a Laplace surface, a two-dimensional surface that represents the density of KBOs in three-dimensional space. Others decided to calculate an invisible plane (called the solar systems invariable plane) that passes through the solar systems center-of-mass as a candidate for the ecliptic plane (an imaginary plane of Earths orbit around the Sun) of the Kuiper Belt. However, these measurements yielded inconsistent results as they predicted different inclinations and longitudes of the plane. Some researchers rejected the Laplace surface, some rejected the invariable plane below certain values of the semimajor axes of the Kuiper Belt, while others concluded that they could not reject either.

Instead of changing the data to fit a model, the authors generated a statistical distribution where the KBOs are modeled to have lower eccentricities than the average, and similar tilts to average KBOs. Then, they use a machine learning-based program that predicts possible outcomes of an uncertain event (known as a Monte Carlo Simulation) to account for bias in calculation. For each of the KBOs in the simulation, these objects are then assigned orbits randomly based on a statistical model of KBOs in real life. The program compares the properties of each simulated project to those of real objects, to ensure that the sample represents the biases. They repeat this experiment and calculate the size of the model each time with a new randomly assigned orbit for each object. This mass random generated list by computers form a predictable pattern, where machines can estimate the most possible positions and orbits of the KBOs, and use this to calculate how likely it is for the plane to be a certain size.

A successful machine learning experiment needs a large dataset. The bigger the dataset is, the more data a computer can use to recognize patterns. Luckily for us, we have the data from JPL Solar System Dynamics Groups Small Body Database Query, which contains almost all asteroid-type objects. With some cross-referencing on different databases, the authors were able to create a dataset that is large enough for us to continue our simulations.

The authors were able to develop a calculation for the size of Kuiper Belt and its uncertainty, based on the machine learning technique discussed in the previous paragraph. Figure 2 is a model of the Kuiper Belt in an area represented by (p,q), parameterised representations of the inclination angle.

In the figure, (a) is the classic measurements,(b) is the entire belt from 35-150 AU, and (c), the entire belt minus part (a). The best-fit plane developed from a series of datasets is shown in dark green +, while the confidence levels are represented by the green circles around. The simulations of the 40000 Monte-Carlo samples are in light green. The invariable plane, proposed by other scientists is represented by the black x, and the theoretical prediction for the Laplace surface is plotted in blue. Since x is well outside the contours, the researchers have concluded that the invariable plane can be rejected because it is too far away from their calculations, but the Laplace surface cannot be rejected as it still lands in the uncertainty intervals based on the simulations.

All in all, this research is able to establish a more accurate measurement for the size of the Kuiper belt. This is a great step in the field of astronomy, and it allows future researchers to develop theories on exoplanets and extraterrestrial life. There is still much more in space waiting for us to discover, and more accurate measurements of our Kuiper Belt is only the beginning.

Astrobite edited by William Lamb

View original post here:
Machine learning, but in space | astrobites - Astrobites

Read More..

Machine learning analysis of research citations highlights … – University of Wisconsin-Madison

Biomedical research aimed at improving human health is particularly reliant on publicly funded basic science, according to a new analysis boosted by artificial intelligence.

B. Ian Hutchins

What we found is that even though research funded by the National Institutes of Health makes up 10% of published scientific literature, those published papers account for about 30% of the substantive research the important contributions supporting even more new scientific findings cited by further clinical research in the same field, says B. Ian Hutchins, a professor in the University of WisconsinMadisons Information School, part of the School of Computer, Data & Information Sciences. Thats a pretty big over-representation.

Hutchins and co-authors Travis Hoppe, now a data scientist at the Centers for Disease Control and Prevention, and UWMadison graduate student Salsabil Arabi, published their findings recently in the Proceedings of the National Academy of Sciences.

Published research papers typically include lengthy sections citing all the previous work supporting or referenced within the study. Predicting substantive biomedical citations without full text, the paper by Hutchins and Hoppe that you are reading about right now, cited no fewer than 64 other studies and sources in its References section.

Citations represent the transfer of knowledge from one scientist (or group of scientists) to another. Citations are extensively catalogued and tracked to measure the significance of individual studies and of the individuals conducting them, but not all citations included in any given paper make equally important contributions to the research they describe.

Were taught that as scientists, when we make a factual claim, were supposed to back it up with some kind of empirical evidence, Hutchins says. Like in Wikipedia entries, you cant have the little citation needed here flag. You have to add that citation. But if that fact youre citing isnt actually describing key prior work that you built upon, then it doesnt really support the interpretation that the citation represents a necessary earlier step toward your results.

Hutchins and his collaborators figured citations added later in the publication process, like those that appear at the behest of peer reviewers the subject-matter experts that evaluate scientific papers submitted to journals are less likely to have been truly important to the authors research.

If youre building on other peoples work, you probably identify that work earlier on in the research process, Hutchins says. That doesnt mean all the references that are in an early version of the manuscript are important ones, but the important ones are probably more concentrated in that earlier version.

To make the early-late distinction, the researchers trained a machine learning algorithm to judge citations on their importance by feeding it citation information from a pool of more than 38,000 scholarly papers. Each papers citation data came in two versions: a preprint version, posted publicly before peer review, and the eventual published version that had undergone peer review.

The algorithm found patterns to help identify the citations that were more likely to be important to each piece of published science. Those results revealed NIH-funded basic biological science appearing in the weightier citations at a rate three times the size of its share of all published research.

Federal funding for basic research is under constant scrutiny from members of the public and congressional leadership, Hutchins says. This gives us some evidence, not just anecdotes, that this kind of basic research funding is really important for stimulating the kind of clinical research treatments and cures for people that Congress tends to be more receptive to funding.

Read more:
Machine learning analysis of research citations highlights ... - University of Wisconsin-Madison

Read More..

Updates on Wasserstein barycenter part11(Machine Learning) | by … – Medium

Author :

Abstract : In this paper, we focus on computational aspects of the Wasserstein barycenter problem. We propose two algorithms to compute Wasserstein barycenters of m discrete measures of size n with accuracy $e$. The first algorithm, based on mirror prox with a specific norm, meets the complexity of celebrated accelerated iterative Bregman projections (IBP), namely $widetilde O(mnsqrt n/e)$, however, with no limitations in contrast to the (accelerated) IBP, which is numerically unstable under small regularization parameter. The second algorithm, based on area-convexity and dual extrapolation, improves the previously best-known convergence rates for the Wasserstein barycenter problem enjoying $widetilde O(mn/e)$ complexity.

2. Distributed Optimization with Quantization for Computing Wasserstein BarycentersarXiv)

Author : Roman Krawtschenko, Csar A. Uribe, Alexander Gasnikov, Pavel Dvurechensky

Abstract : We study the problem of the decentralized computation of entropy-regularized semi-discrete Wasserstein barycenters over a network. Building upon recent primal-dual approaches, we propose a sampling gradient quantization scheme that allows efficient communication and computation of approximate barycenters where the factor distributions are stored distributedly on arbitrary networks. The communication and algorithmic complexity of the proposed algorithm are shown, with explicit dependency on the size of the support, the number of distributions, and the desired accuracy. Numerical results validate our algorithmic analysis.

Go here to see the original:
Updates on Wasserstein barycenter part11(Machine Learning) | by ... - Medium

Read More..

Machine learning-based analysis of regional differences in out-of … – Nature.com

Study design

We conducted a retrospective study utilizing prospectively recorded Japanese Utstein-style EMS activity records. The Ethics Committee of Nara Medical University approved the study (No. 3353), and the requirement for informed consent was waived owing to the use of anonymized records. This study was conducted in accordance with the tenets of the Declaration of Helsinki.

Japan has an aging population as 28.9% of its 130 million people are aged>65years10. The country consists of 47 prefectures with varying population densities of 65.46,399.5 individuals/km2. EMSs respond to all emergency calls and transport approximately 125,000 patients with OHCA to hospitals annually11. Emergency protocols, based on the Japanese Resuscitation Councils Resuscitation Guidelines12 and revised every 5years, are developed and implemented by 250 regional health managers. Each medical control region is supervised by a council established in each prefecture, tailoring protocols to local conditions13,14,15. EMS activities are recorded in the Utstein style and verified by the medical control council, and all records are collected annually by the Fire and Disaster Management Agency11. Our analysis included prehospital records of patients with OHCA resuscitated by EMS and transported to hospitals in 47 prefectures between 2015 and 2020, excluding patients aged<18years and those with non-cardiogenic cardiopulmonary arrest to reduce pathology variability.

In Japan, EMS is activated via a Communications Command Center upon receiving emergency calls. Bystanders may be instructed to administer cardiopulmonary resuscitation (CPR) over the telephone if cardiac arrest is suspected. Each ambulance includes a team of three, often featuring emergency life-saving technicians capable of advanced airway management and adrenaline administration for OHCA, under online medical control supervision. Additionally, hospital destinations are determined during field operations, and all patients, barring those with evident signs of death, are transported to a hospital.

We employed 23 factors and prefecture numbers from the Utstein-style EMS activity records as predictors, including county number, age, year and month of onset, bystander type, initial rhythm, number of defibrillations, number of adrenaline boluses administered, and elapsed time of each activity. Notably, the prefecture number was treated as a continuous variable due to its sequential allocation from north to south. This approach aimed to capture potential spatial correlations between adjacent prefectures. We also conducted a similar analysis using one-hot encoding for the prefecture numbers, and the outcomes did not contradict the results obtained when treating the prefecture number as a continuous variable. Categorical data were one-hot encoded. Remarkably, in the case of missing data, we refrained from substituting them with any particular value. Instead, the data missingness was coded as a separate category, which was incorporated into our analysis as a separate data element. Selected continuous variables were standardized using z-score normalization, a method that confers advantages in machine learning algorithms such as neural networks by aiding gradient descent convergence and mitigating issues related to weight initialization and gradient problems. Time factors, which were initially considered continuous variables, were one-hot encoded as categorical data16 because of their non-linear relationship with prognosis in cardiopulmonary resuscitation. The time factors were measured in minutes and thus represented as 1, 2, 3, 4, minutes.

Cases in which a specific intervention, such as defibrillation or drug administration, was not performed were also considered. These were coded as no intervention and incorporated into the contact-to-intervention column, allowing the model to reflect a comprehensive range of patient experiences. These steps resulted in 249 features (see Supplementary Table S1). Subsequently, we constructed a machine learning model to predict good neurological outcomes 1month after cardiac arrest, based on the cerebral performance category (CPC) score17a binary classification (Yes/No), with CPC1/2 signifying good neurological outcome and CPC3-5 indicating poor neurological outcomesourced from the Utstein records.

We stratified and randomly split the training and test datasets using an 8:2 ratio based on CPC1/2 to ensure a consistent ratio for predictive model construction. The prediction model was built using the neural network with the best average class sensitivity after several machine learning model trials. The compared methods included logistic regression, support vector machine, decision tree, random forest, and LightGBM9. To balance model bias (underfitting) and variance (overfitting), we applied a stratified cross-validation method (five-fold) using CPC1/2, along with batch normalization and dropouts in each neural network layer. The models accuracy plateaued after increasing the number of layers to five because of which we used a five-layer network to optimize learning costs. The sigmoid function served as the activation function and binary cross-entropy served as the loss function18. We measured model performance using area under the receiver operating characteristic curve (AUROC) and accuracy during training.

Imbalanced datasets significantly affect minority class performance. To address misclassification, we simulated based on predicted CPC1/2 numbers and employed class weighting during training to balance sensitivities, considering trade-offs. Our model aimed to maximize the majority class (CPC35) sensitivity without excessively reducing minority class (CPC1/2) sensitivity. We set CPC1/2 sensitivity at 80% and tested weights from 1 to 100 in 0.1 increments to optimize CPC3-5 sensitivity.

Additional training parameters included a batch size of 1,024,100 epochs, a learning rate of 0.001, and Adam optimizer. We conducted training using Python version 3.8.5 (Python Software Foundation, Beaverton, OR, USA).

We assessed the association of EMS activity duration with predicted CPC1/2 counts by simulating the constructed prediction model on a test dataset (n=92,108), containing all previously split prefectures from the training set. The simulation methodology involved three time factors: elapsed time from EMS arrival to hospital arrival (a), EMS arrival to first defibrillation (b), and EMS arrival to first drug administration (c).

Previous studies have shown that these temporal factors are important prognostic predictors of EMS activity time19,20,21,22,23,24,25,26. For example, shorter time from EMS arrival to defibrillation19,25 and from EMS arrival to drug administration20,21,22,23,24,25 are associated with better survival and improved neurological outcomes in OHCA patients. The prognostic impact of EMS providers staying on scene and performing their activities has also been reported26. Patients with non-shockable initial rhythm were excluded for (b), and those with EMS-witnessed cardiac arrest were excluded for (c). Time factors increased or decreased by5 to+5min for defibrillation and drug administration, and from5 to+10min for EMS arrival to hospital arrival time, in 1-min increments. We created a dataset adjusting each time factor in the test dataset and calculated the average predicted CPC1/2 score using the created prediction model. Then, we determined the percentage change in mean predicted CPC1/2 count to assess the association of time increase/decrease with the unadjusted data. We focused on percentage change relative to unadjusted data for a prefecture-specific analysis. A heat map visualized and evaluated the proportion of change between time adjustment and mean predicted CPC1/2 count.

We employed the same time adjustment method to estimate and visualize predicted CPC1/2 counts for the test dataset split by prefecture. We identified the time adjustments most associated with prognosis in each prefecture for the combinations (a) & (b) and (a) & (c), revealing treatment and EMS arrival to hospital arrival time adjustments with the greatest potential to improve predicted prognosis.

Patient characteristics are summarized as medians and interquartile ranges (IQRs) for continuous variables and counts and percentages for categorical variables. Additionally, the evaluation metric for the five models is expressed as meansstandard deviations. The standard deviations were calculated based on the variations in the evaluation metric across the five-fold cross-validation.

See the original post:
Machine learning-based analysis of regional differences in out-of ... - Nature.com

Read More..

Dr. Kavoussi discusses machine learning and AI-based tools’ roles … – Urology Times

In this interview, Nicholas L. Kavoussi, MD, discusses his clinical practice and research interests pertaining to kidney stones. Kavoussi is an assistant professor of urology at Vanderbilt University Medical Center in Nashville, Tennessee.

My clinical practice is in minimally invasive treatment of benign prostate disease and kidney stone disease, mainly through natural orifice surgery. My research interest is in developing novel tools to improve endoscopic surgery of kidney stone disease.

I think the biggest unmet need is that kidney stone disease is common, and after you have a kidney stone, recurrence is common. We don't have a great way of identifying specific risk factors, and adequately mitigating those risk factors to prevent recurrence events and stone formation. The hope is by doing some of the work here, where we're trying to improve our endoscopic surgical ability using novel tools, we can more accurately treat stones and prevent future stone recurrence episodes.

We have research support from the NIH through a novel technology R21 grant. The intent of this project is to improve endoscopic visibility and navigation during kidney stone surgery. So we're building maps and tracking kidney stones in low visibility settings. The goal of this project is to be able to teach computers what we see when we operate. That way, the computer can know what surgery is supposed to look like and build these maps and track kidney stones while we operate to make us more accurate and track potential harmful fragments while we treat these kidney stones.

I think these machine learning and artificial intelligence-based [tools] will impact the way we practice clinically, especially with chronic kidney stone disease. I think our biggest issue right now is we don't have the numbers and the datasets needed to really evaluate these technologies and how they might impact our lives and the lives of our patients. So I think the role of these technologies, though they're still in their infancy and have a long way to go, will be pretty impactful in terms of how we treat and take care of patients kidney stone.

I think the key to remember for a lot of what we're doing is that kidney stone disease is complicated. It's a really heterogenous patient population. There are a lot of reasons why stones form. There are a lot of reasons why they recur; we don't quite understand that. I think the way we work up stones now is very variable from provider to provider, and really needs to fit patient's needs rather than our clinical workflow. My hope is that these machine learning-based tools and computer vision models will really allow for patient-directed, specific care. And in terms of how we treat the stone surgically and the tools we're building, this is really 1 facet of stone care. And I think really, it's important to consider all different factors that contribute to stone disease and recurrence when helping these patients.

This transcript was edited for clarity.

Originally posted here:
Dr. Kavoussi discusses machine learning and AI-based tools' roles ... - Urology Times

Read More..

Azerbaijani young talent: From computer science to AI – The Jerusalem Post

Introduction

We are pleased to present Mehdi Rasul, a talented teenager excelling in his 11th-grade studies within the A-Level program at LANDAU School in Azerbaijan. Since a young age, he has been interested in Artificial Intelligence and Machine Learning, especially in Natural Language Processing and deep learning technologies. Over the years, Mehdi has become proficient in programming languages such as Python, Numpy, and Javascript. His exceptional academic performance is a testament to his unwavering quest for knowledge. Beyond the classroom, Mehdi's accomplishments extend to a diverse range of extracurricular activities, including community service, internships in various AI-related projects, active participation in MUN conferences, membership in the school newsletter, the completion of the IT Essentials program at Cisco Networking Academy, etc. In 2021 and 2022, Mr. Rasul won the competitions hosted by the "CyberMath Academy" and in 2022, he made a notable appearance in the American Mathematics Olympiad. Mehdi's research project has already been featured on different platforms and published in esteemed scientific journals in both Azerbaijan and Switzerland. His research delves into the comparison of diverse machine learning algorithms for sentiment analysis in both Azerbaijani and English texts. The study's findings underscore the pivotal role of national language corpora in achieving precise outcomes in natural language processing, particularly in sentiment analysis.

The article is provided below for your perusal.

The prediction of the sentiment of the text within different business spheres has been a challenging problem for various languages. Sentiment analysis, also known as opinion mining, is an active area of research in natural language processing (NLP) and computational linguistics that involves using text analysis and classification methods to identify and extract subjective information such as opinions, emotions, and attitudes from text data. Sentiment analysis has many applications in business, politics, and social media monitoring. Using sentimental analysis of textual data, companies can evaluate customer feedback, monitor reputation, forecast future and user behavior, etc., which can lead to better performance, efficiency, and increased profits.CREDIT: Mehdi Rasul

Studies have shown that machine learning algorithms, particularly those using supervised learning and deep learning, produce satisfactory results in the automation of textual sentiment analysis. Recent research illustrates that sentiment analysis of given texts can be automated, relying on the combination of computer science and mathematics, while also improving accuracy.

A well-defined English language corpus for model training helps build powerful and highly accurate models for English texts. Given the extensive array of libraries offering diverse natural language processing techniques in English, an automated system for analyzing texts and extracting meaningful insights or summaries from input paragraphs is relatively easy to create. The abundance of useful libraries in computer linguistics for English, combined with numerous natural language processing tools, such as word correction, grammar checking, text generation, word tokenization, etc., enables the construction of highly accurate sentiment analysis models.

Despite its utilization with major languages such as English, research on sentiment analysis of under-resourced languages like Azerbaijani is still relatively limited. Azerbaijani is the official language of the Republic of Azerbaijan and has distinct linguistic characteristics. However, it still lacks labeled datasets and the sophisticated language corpus for training machine learning models in sentiment analysis of the texts. Therefore, building sentiment analysis for Azerbaijani appears to be uniquely challenging, due to the lack of built-in NLP techniques and language corpus designed specifically for Azerbaijani.

This study encompasses a comprehensive comparison of machine learning algorithms. These algorithms are evaluated to compare their effectiveness in achieving sentiment classification for Azerbaijani. The same algorithms were tested on the English version of the dataset to address the importance of the built-in language corpus and the advancement of currently applicable techniques for NLP tasks in Azerbaijani.

Recent studies have shown that using machine learning techniques, such as supervised and deep learning algorithms, has contributed to significant advancements in improving and automating sentiment analysis for textual data. In politics, machine learning is also used to determine the sentiment of texts. In 2016, Heredia et al. collected political tweets to predict the U.S. 2016 election. Researchers collected three million location-based tweets related to Donald Trump and Hillary Clinton and trained them on the deep convolutional neural network (CNN) to predict the election results and attained an accuracy score of 84%.

The sentiment analysis of the texts in Azerbaijani is limited due to the lack of the labeled dataset and the language corpus. In 2013, research by Neethu and Rajasree on sentiment analysis of Twitter using machine learning algorithms produced satisfactory results in terms of classification of tweets into positive and negative classes.

Generally, machine learning algorithms perform well in the classification of texts in both Azerbaijani and English. The major algorithms utilized in text classification problems are Support Vector Machines, Nave Bayes, and Logistic Regression, which help to identify the patterns in both type and sentiments. Decision trees are supervised machine-learning algorithms used for classification and regression tasks. The goal is to create a model that predicts the value of a target variable by using decision rules created during the training process from the input features.

In this research, various machine learning algorithms were used to predict the sentiment of movie reviews in both English and Azerbaijani. The study has indicated the achievements attained in building models with various techniques. TF-IDF and BOW (or Count Vectorizer) have been implemented for feature extraction methods from the texts and tested in different models, including Logistic Regression, Nave Bayes, SVM, Decision Tree, Random Forest, AdaBoost, and XGBoost.

For the Azerbaijani version of the dataset, using the TF-IDF feature extraction approach, Logistic Regression and SVM algorithms produced better results compared to other models. However, the decision tree had 64% accuracy. When BOW feature extraction was used, Logistic Regression produced 84% accuracy, while SVM reached an accuracy of 82%.

The same dataset in English has also been modeled to compare the results with different preprocessing techniques that were not applied in Azerbaijani, such as stemming, stopwords list, etc. Overfitting was much less of an issue for models based on English datasets. The highest score was attained with the BOW feature extraction method using the Logistic Regression method. Additionally, SVM and Naive Bayes algorithms performed well with the TF-IDF feature extraction method and achieved an 85% accuracy score. The research indicates that the results are similar for both language models.

Python has enriched libraries for building language models in English and includes a list of stopwords, word tokenization, stemming, and lemmatization techniques. The English language corpus is well developed, which helps achieve higher results, unlike Azerbaijani, which lacks the language corpus, making it harder to build generalized models. Additional resources should be dedicated to research and creating a language corpus for Azerbaijani, which would help attain greater accuracy.

This article was written in cooperation with Mehdi Rasul

Visit link:
Azerbaijani young talent: From computer science to AI - The Jerusalem Post

Read More..