Page 587«..1020..586587588589..600610..»

Gemini AI: What do we know about Google’s answer to ChatGPT? – Livescience.com

Google DeepMind has released a rival to ChatGPT, named Gemini, and it can understand and generate multiple types of media including images, videos, audio, and text.

Most artificial intelligence (AI) tools only understand and generate one type of content. For example, OpenAI's ChatGPT, "reads" and creates only text. But Gemini can generate multiple types of output based on any form of input, Google said in a blog post.

The three versions of Gemini 1.0 are Gemini Ultra, the largest version, Gemini Pro, which is being rolled out into Google's digital services, and Gemini Nano, designed to be used on devices like smartphones.

According to DeepMind's technical report on the chatbot, Gemini Ultra beat GPT-4 and other leading AI models in 30 of 32 key academic benchmarks used in AI research and development. These include high school exams and tests on morality and law.

Specifically, Gemini won out in nine image comprehension benchmarks, six video understanding tests, five in speech recognition and translation, and 10 of 12 text and reasoning benchmarks. The two in which Gemini Ulta failed to beat GPT-4 were in common-sense reasoning, according to the report.

Related: AI is transforming every aspect of science. Here's how.

Building models that process multiple forms of media is hard because biases in the training data are likely to be amplified, performance tends to drop significantly, and models tend to overfit meaning they perform well when tested against the training data, but can't perform when exposed to new input.

Multimodal training also normally involves training different components of a model separately, each on a single type of medium and then stitching these components together. But Gemini was trained jointly across text, image, audio and video data at the same time. Scientists sourced this data from web documents, books and code.

Scientists trained Gemini by curating the training data and incorporating human supervision in the feedback process.

The team deployed servers across multiple data centers on a much grander scale than previous AI training efforts and relied on thousands of Google's AI accelerator chips known as the tensor processing units (TPUs).

DeepMind built these chips specifically to speed up model training, and DeepMind packaged them into clusters of 4,096 chips known as "SuperPods", before training its system. The overall result of the re-configured infrastructure and methods meant the goodput the volume of genuinely useful data that moved through the system (as opposed to throughput, which is all data) increased from 85% in previous training endeavors to 97%, according to the technical report.

DeepMind scientists envision the technology being used in scenarios such as a person uploading photos of a meal being prepared in real-time, and Gemini responding with instructions on the next step in the process.

That said, the scientists did concede hallucinations a phenomenon in which AI models return false information with maximum confidence remains an issue for Gemini. Hallucinations are normally caused by limitations or biases in the training data, and they're difficult to eradicate.

Link:
Gemini AI: What do we know about Google's answer to ChatGPT? - Livescience.com

Read More..

Advancements in machine learning for machine learning Google Research Blog – Google Research

Posted by Phitchaya Mangpo Phothilimthana, Staff Research Scientist, Google DeepMind, and Bryan Perozzi, Senior Staff Research Scientist, Google Research

With the recent and accelerated advances in machine learning (ML), machines can understand natural language, engage in conversations, draw images, create videos and more. Modern ML models are programmed and trained using ML programming frameworks, such as TensorFlow, JAX, PyTorch, among many others. These libraries provide high-level instructions to ML practitioners, such as linear algebra operations (e.g., matrix multiplication, convolution, etc.) and neural network layers (e.g., 2D convolution layers, transformer layers). Importantly, practitioners need not worry about how to make their models run efficiently on hardware because an ML framework will automatically optimize the user's model through an underlying compiler. The efficiency of the ML workload, thus, depends on how good the compiler is. A compiler typically relies on heuristics to solve complex optimization problems, often resulting in suboptimal performance.

In this blog post, we present exciting advancements in ML for ML. In particular, we show how we use ML to improve efficiency of ML workloads! Prior works, both internal and external, have shown that we can use ML to improve performance of ML programs by selecting better ML compiler decisions. Although there exist a few datasets for program performance prediction, they target small sub-programs, such as basic blocks or kernels. We introduce TpuGraphs: A Performance Prediction Dataset on Large Tensor Computational Graphs (presented at NeurIPS 2023), which we recently released to fuel more research in ML for program optimization. We hosted a Kaggle competition on the dataset, which recently completed with 792 participants on 616 teams from 66 countries. Furthermore, in Learning Large Graph Property Prediction via Graph Segment Training, we cover a novel method to scale graph neural network (GNN) training to handle large programs represented as graphs. The technique both enables training arbitrarily large graphs on a device with limited memory capacity and improves generalization of the model.

ML compilers are software routines that convert user-written programs (here, mathematical instructions provided by libraries such as TensorFlow) to executables (instructions to execute on the actual hardware). An ML program can be represented as a computation graph, where a node represents a tensor operation (such as matrix multiplication), and an edge represents a tensor flowing from one node to another. ML compilers have to solve many complex optimization problems, including graph-level and kernel-level optimizations. A graph-level optimization requires the context of the entire graph to make optimal decisions and transforms the entire graph accordingly. A kernel-level optimization transforms one kernel (a fused subgraph) at a time, independently of other kernels.

To provide a concrete example, imagine a matrix (2D tensor):

It can be stored in computer memory as [A B C a b c] or [A a B b C c], known as row- and column-major memory layout, respectively. One important ML compiler optimization is to assign memory layouts to all intermediate tensors in the program. The figure below shows two different layout configurations for the same program. Lets assume that on the left-hand side, the assigned layouts (in red) are the most efficient option for each individual operator. However, this layout configuration requires the compiler to insert a copy operation to transform the memory layout between the add and convolution operations. On the other hand, the right-hand side configuration might be less efficient for each individual operator, but it doesnt require the additional memory transformation. The layout assignment optimization has to trade off between local computation efficiency and layout transformation overhead.

If the compiler makes optimal choices, significant speedups can be made. For example, we have seen up to a 32% speedup when choosing an optimal layout configuration over the default compilers configuration in the XLA benchmark suite.

Given the above, we aim to improve ML model efficiency by improving the ML compiler. Specifically, it can be very effective to equip the compiler with a learned cost model that takes in an input program and compiler configuration and then outputs the predicted runtime of the program.

With this motivation, we release TpuGraphs, a dataset for learning cost models for programs running on Googles custom Tensor Processing Units (TPUs). The dataset targets two XLA compiler configurations: layout (generalization of row- and column-major ordering, from matrices, to higher dimension tensors) and tiling (configurations of tile sizes). We provide download instructions and starter code on the TpuGraphs GitHub. Each example in the dataset contains a computational graph of an ML workload, a compilation configuration, and the execution time of the graph when compiled with the configuration. The graphs in the dataset are collected from open-source ML programs, featuring popular model architectures, e.g., ResNet, EfficientNet, Mask R-CNN, and Transformer. The dataset provides 25 more graphs than the largest (earlier) graph property prediction dataset (with comparable graph sizes), and graph size is 770 larger on average compared to existing performance prediction datasets on ML programs. With this greatly expanded scale, for the first time we can explore the graph-level prediction task on large graphs, which is subject to challenges such as scalability, training efficiency, and model quality.

We provide baseline learned cost models with our dataset (architecture shown below). Our baseline models are based on a GNN since the input program is represented as a graph. Node features, shown in blue below, consist of two parts. The first part is an opcode id, the most important information of a node, which indicates the type of tensor operation. Our baseline models, thus, map an opcode id to an opcode embedding via an embedding lookup table. The opcode embedding is then concatenated with the second part, the rest of the node features, as inputs to a GNN. We combine the node embeddings produced by the GNN to create the fixed-size embedding of the graph using a simple graph pooling reduction (i.e., sum and mean). The resulting graph embedding is then linearly transformed into the final scalar output by a feedforward layer.

Furthermore we present Graph Segment Training (GST), a method for scaling GNN training to handle large graphs on a device with limited memory capacity in cases where the prediction task is on the entire-graph (i.e., graph-level prediction). Unlike scaling training for node- or edge-level prediction, scaling for graph-level prediction is understudied but crucial to our domain, as computation graphs can contain hundreds of thousands of nodes. In a typical GNN training (Full Graph Training, on the left below), a GNN model is trained using an entire graph, meaning all nodes and edges of the graph are used to compute gradients. For large graphs, this might be computationally infeasible. In GST, each large graph is partitioned into smaller segments, and a random subset of segments is selected to update the model; embeddings for the remaining segments are produced without saving their intermediate activations (to avoid consuming memory). The embeddings of all segments are then combined to generate an embedding for the original large graph, which is then used for prediction. In addition, we introduce the historical embedding table to efficiently obtain graph segments embeddings and segment dropout to mitigate the staleness from historical embeddings. Together, our complete method speeds up the end-to-end training time by 3.

Finally, we ran the Fast or Slow? Predict AI Model Runtime competition over the TpuGraph dataset. This competition ended with 792 participants on 616 teams. We had 10507 submissions from 66 countries. For 153 users (including 47 in the top 100), this was their first competition. We learned many interesting new techniques employed by the participating teams, such as:

We will debrief the competition and preview the winning solutions at the competition session at the ML for Systems workshop at NeurIPS on December 16, 2023. Finally, congratulations to all the winners and thank you for your contributions to advancing research in ML for systems!

If you are interested in more research about structured data and artificial intelligence, we hosted the NeurIPS Expo panel Graph Learning Meets Artificial Intelligence on December 9, which covered advancing learned cost models and more!

Sami Abu-el-Haija (Google Research) contributed significantly to this work and write-up. The research in this post describes joint work with many additional collaborators including Mike Burrows, Kaidi Cao, Bahare Fatemi, Jure Leskovec, Charith Mendis, Dustin Zelle, and Yanqi Zhou.

Link:
Advancements in machine learning for machine learning Google Research Blog - Google Research

Read More..

From hallucinations to discovery: For the first time a large language model finds new solutions to math problems – ZME Science

Credit: DALL-E 3.

Large Language Models (LLMs) like ChatGPT have a lot of things going for them. These powerful AI systems can synthesize and interpret vast amounts of information and are surprisingly human-like with language. At the same time, theyre also notorious for making up facts with confidence. Put simply, they hallucinate, as people have come to describe this annoying behavior.

A huge question ever since this technology was released is whether LLMs are capable of discovering new knowledge, rather than repurposing and rehashing existing information. As it turns out, they can.

Researchers at Googles DeepMind branch have shown a new AI method called FunSearch, which can forge new paths to find solutions to complex problems in mathematics and computer science.

The innovation of FunSearch lies in the pairing of a pre-trained LLM with an automated evaluator. This setup is designed to leverage the LLMs strength in generating creative solutions in the form of computer code, while the evaluator rigorously checks these solutions for accuracy. The highest-performing solutions are continuously fed back into the cycle, fostering a self-improving loop of problem-solving and innovation.

This partnership enables an iterative refinement process, transforming initial creative outputs into verified, novel knowledge. The focus on discovering functions in computer code is what gives FunSearch its distinctive name and operational approach.

This initiative marks the first time LLMs have contributed to solving open problems in the scientific and mathematical community. FunSearch found novel solutions to the cap set problem, a long-standing mathematical challenge.

The Cap Set Problem in mathematics involves finding the largest subset of integers from 0 to 3n1 (where each integer is represented in base 3) such that no three integers in the subset sum to another integer in base 3. Its a challenge in combinatorics, a field concerned with counting, arrangement, and structure. Terence Tao, the highest IQ person in the world and one of the worlds leading mathematicians, once described the cap set problem as one of his favorite open questions in the field.

FunSearch succeeded in discovering new, larger cap sets, contributing valuable insights to the problem and demonstrating the potential of AI in advancing mathematical research. FunSearchs contribution marks the largest increase in the size of cap sets in the past two decades.

These results demonstrate that the FunSearch technique can take us beyond established results on hard combinatorial problems, where intuition can be difficult to build. We expect this approach to play a role in new discoveries for similar theoretical problems in combinatorics, and in the future it may open up new possibilities in fields such as communication theory, wrote the DeepMind researchers in a blog post.

Moreover, FunSearch has proven itself further by enhancing algorithms for the bin-packing problem. The bin-packing problem is a classic algorithmic challenge. It involves efficiently packing objects of different sizes into a finite number of bins or containers in a way that minimizes the number of bins used.

Contrary to many computational tools that offer solutions without explanation like a black box, FunSearch provides a detailed account of how its conclusions are reached.

This show-your-working approach is how scientists generally operate, with new discoveries or phenomena explained through the process used to produce them, add the DeepMind researchers.

The ability of FunSearch to not only generate innovative solutions but also provide the details of the problem-solving process holds immense potential. With the continual advancement of LLM technology, the capabilities of tools like FunSearch are expected to expand, paving the way for groundbreaking discoveries and solutions to some of societys most pressing scientific and engineering challenges.

The findings were reported in the journal Nature.

View post:
From hallucinations to discovery: For the first time a large language model finds new solutions to math problems - ZME Science

Read More..

Using geospatial data & machine learning to predict water quality in Ethiopia – World Bank

Simone D. McCourtie / The World Bank

Expanding access to safe drinking water in low and middle-income countries is a key human development priority, with targets set at national and global levels. As such, ensuring access to safe water, sanitation and hygiene for all is also one of the Sustainable Development Goals.

Information on drinking water quality is key to monitoring progress towards achieving global and national targets. Therefore, it is important to accurately measure quality of drinking water accessed by households and individuals and to determine if drinking water is free from biological and chemical contamination.

An increasing number of countries (over 50 to date) have now integrated objective water quality testing in national household surveys to monitor access to safely managed drinking water services. This approach enables the collection of representative information for general household populations, with the potential to disaggregate results by different geographic and socioeconomic groups.

The ability to link water quality information to the wealth of information collected in household surveys facilitates research, as well as the identification of effective interventions to improve access to safely managed drinking water services.

However, integrating water quality testing in household surveys requires additional financial resources and specialized technical assistance, and can increase burden on statistical agencies, especially in resource-constrained contexts. For example, E. coli testing in the field requires equipment, consumables and dedicated training for field staff on aseptic techniques, incubation and interpreting results.

In our recent study, Addressing gaps in data on drinking water quality through data integration and machine learning: evidence from Ethiopiaa collaborative work between the World Bank Living Standards Measurement Study (LSMS) team and the Joint Monitoring Programme (JMP) of the World Health Organization and UNICEFwe proposed an approach to fill data gaps in drinking water quality.

Lets unwrap our methodology step by step.

The idea is that while it may not be logistically and financially possible to implement water quality testing in every household survey, data obtained from a recent survey can be integrated with publicly available geospatial data on rainfall, temperature, proximity to the nearest market and roads, among others, and in turn used to train a machine learning model to generate reliable insights on drinking water quality in years when no surveys were conducting tests on the ground.

The country selected as a case study for our research was Ethiopia. In 2016, when the latest data on water quality was collected as part of the Ethiopian Socioeconomic Survey, about 68 percent of households had access to drinking water from improved sources, such as piped sources, protected wells and springs. However, over half of those improved sources were contaminated (Figure 1).

Using water quality testing data from the third wave of the Ethiopia Socioeconomic Survey (ESS3) in 2016, our study, Addressing gaps in data on drinking water quality through data integration and machine learning: evidence from Ethiopia, examined the performance of a range of commonly used machine learning algorithms to predict E. coli contamination in the households drinking water sources.

The study developed a predictive model for contamination of drinking water sources by integrating socioeconomic survey data with geospatial data sources on the basis of household GPS locations. It compared a few commonly used classification algorithms including GLM, GLMNET, KNN, Support Vector Machine, and two decision tree-based classifiers: Random Forest (RF), and XGBoost. RF performed the best across most metrics, with XGBoost becoming a close runner up.

The study also examined the performance of different groups of predictors variables, namely household demographic and socioeconomic attributes, water service particularities and geospatial variables, on the performance of the algorithms and applied the predictive models to other waves of the ESS, in 2013/14 and 2018/19.

Overall, predictions for ESS3 (2015/16 ESS) were comparable to the actual data under different scenarios. The study finds that a model that has all prospective predictor variables is found to have a strong discrimination ability (Area under the curve (AUC) 0.91; 95% Confidence Interval (CI) 0.89, 0.94). Model performance was poor when type of water source was the only predictor (AUC 0.80; 95% CI 0.77, 0.84).

However, augmenting water source variables with selected household-level socioeconomic predictors, but excluding geospatial variables, resulted in a performance comparable to the full model (AUC 0.89; 95% CI 0.86, 0.91).

The model with only geospatial predictor also achieved a performance that was comparable to the full model (AUC 0.91; 95% CI 0.88, 0.93). The geospatial variables are also key predictors of contamination in the full predictive model (Figure 2).

Overall, three key take-away messages emerge from our study:

Machine learning approaches can be used to develop a model and fill the gap that might arise due to the challenges of implementing a water quality testing.

A georeferenced household survey with objective water testing and basic data on socioeconomic attributes, integrated with geospatial data sources, can be used to develop reliable predictive models for drinking water quality.

Provided that the data from a recent survey with objective water quality testing exist, predictive machine learning models relying exclusively on geospatial variables may also suffice for understanding variations in risk of E. coli contamination in drinking water sources and generating water quality risk maps.

Visit link:
Using geospatial data & machine learning to predict water quality in Ethiopia - World Bank

Read More..

Using deep learning to identify teens most in need of mental health support – Medical Xpress

This article has been reviewed according to ScienceX's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

trusted source

proofread

close

The personal yet global struggle with mental health may be more visible now than ever before. Yet many people still find it difficult to access the support they need.

In Japan, suicide is sadly the leading cause of death for young people. Researchers, including from the University of Tokyo, have carried out a six-year study to better understand the myriad factors which can impact adolescent mental health. After surveying 2,344 adolescents and their caregivers, and using computer-based deep learning to process the results, they were able to identify five categories into which the young people could be grouped.

Nearly 40% of those involved were classified as groups with some problems. Of these, almost 10% lived with mental health problems that had not been identified by their caregivers. This group was most at risk of self-harm and suicidal ideation. Identifying the factors that may lead young people to suicide and who is most at risk is key to supporting preventive efforts and early intervention.

Last year in Japan, 514 youths and children aged 18 and younger tragically lost their lives to suicide. This was the highest number for this age group since records began in 1978. Suicide is the leading cause of death for people aged 15 to 34 years old, according to data from Japan's Ministry of Health, Labor and Welfare. While adult suicide rates have been generally declining over the past 10 to 15 years, the reverse has been noted for adolescents. Officials speculate that school-related issues, difficult personal and family relationships, and lingering impacts of the pandemic may have contributed to the high number of deaths.

The World Health Organization (WHO) identifies suicide as a major global public health concern, but also says it is preventable through evidence-based interventions and by addressing factors that can lead to poor mental health. Researchers from the University of Tokyo and the Tokyo Metropolitan Institute of Medical Science are analyzing data on various problems in adolescence which were assessed both by self and caregivers, resulting in identification of young people who may be at suicide-related risk.

"We recently found that adolescents who were considered to have no problems by their caregivers actually had the highest suicide-related risk," said Daiki Nagaoka, a doctoral student in the Department of Neuropsychiatry at the University of Tokyo and a hospital psychiatrist. "So it is important that society as a whole, rather than solely relying on caregivers, takes an active role in recognizing and supporting adolescents who have difficulty in seeking help and whose distress is often overlooked."

The team surveyed adolescents and their caregivers in Tokyo over a period of six years. The participants completed self-report questionnaires, answering questions on psychological and behavioral problems such as depression, anxiety, self-harm and inattention, as well as their feelings about family and school life. The team also made note of factors such as maternal health during pregnancy, involvement in bullying and the caregivers' psychological states.

Now published in The Lancet Regional HealthWestern Pacific, the study began when the children were 10 years old, and checked in with them again at ages 12, 14 and 16. Overall, 3,171 adolescents took part, with 2,344 pairs of adolescents and their caregivers participating throughout the full study.

close

"Psychiatry faces challenges in understanding adolescent psychopathology, which is diverse and dynamic. Previous studies typically classified adolescents' psychopathological development based on the trajectories of only two or three indicators. By contrast, our approach enabled the classification of adolescents based on a number of symptom trajectories simultaneously by employing deep-learning techniques which facilitated a more comprehensive understanding," explained Nagaoka.

Deep learning, a computer program that mimics the learning process of our brains, enabled the team to analyze the large amounts of data they collected to find patterns in the responses. By grouping the trajectories of the psychological and behavioral problems identified in the survey, they could classify the adolescents into five groups, which they named based on their key characteristic: unaffected, internalizing, discrepant, externalizing and severe.

The largest group, at 60.5% of the 2,344 adolescents, was made up of young people who were classified as "unaffected" by suicidal behavior.

The remaining 40% were found to be negatively affected in some way. The "internalizing" group (16.2%) persistently internalized problems and showed depressive symptoms, anxiety and withdrawal. The "discrepant" group (9.9%) experienced depressive symptoms and "psychotic-like experiences," but had not been recognized as having such problems by their caregivers. The "externalizing" group (9.6%) displayed hyperactivity, inattention and/or behavioral issues but few other problems.

Finally, the smallest group was categorized as "severe" (3.9%) and dealt with chronic difficulties of which their caregivers were aware, in particular psychotic-like experiences and obsessive-compulsive behavior.

Of all the groups, young people in the "discrepant" category were at highest risk of self-harm and suicidal thoughts. The researchers found that they could significantly predict who would be included in this group based on whether the child avoided seeking help for depression, and whether their caregiver also had a mental health problem.

The researchers suggest that the caregiver's mental state could impact the adolescent's mental health through both genetic factors and parenting environment, such as the caregiver's ability to pay attention to the difficulties an adolescent might face. Although this research has several limitations, it still enabled the team to identify a number of risk factors that could be used to predict which groups adolescents might fall into.

close

"In daily practice as a psychiatrist, I observed that existing diagnostic criteria often did not adequately address the diverse and fluid difficulties experienced by adolescents," said Nagaoka. "We aimed to better understand these difficulties so that appropriate support can be provided. Next we want to better understand how adolescents' psychopathological problems interact and change with the people and environment around them. Recognizing that numerous adolescents face challenges and serious issues, yet hesitate to seek help, we must establish supportive systems and structures as a society."

More information: Identify adolescents' help-seeking intention on suicide through self- and caregiver's assessments of psychobehavioral problems: deep clustering of the Tokyo TEEN Cohort study., The Lancet Regional HealthWestern Pacific (2023). DOI: 10.1016/j.lanwpc.2023.100979

If you or someone you know is struggling, free help and support is available. For a list of helplines around the world, please visit: http://www.suicide.org/international-suicide-hotlines.html

Befrienders International provides confidential support to people in emotional distress or crisis: https://www.befrienders.org

See the rest here:
Using deep learning to identify teens most in need of mental health support - Medical Xpress

Read More..

Featureform Raises $5.5M by Refining and Accelerating the Way Teams Work on AI and ML – Datanami

SAN FRANCISCO, Dec. 15, 2023 Featureform has announced $5.5 million in seed funding led by GreatPoint Ventures and Zetta Venture Partners with participation from Tuesday Capital and Alumni Ventures. This round of capital will allow Featureform to expand its product growth and increase support for existing and new enterprise customers, in addition to its open-source community. The completion of the Seed round brings Featureforms total funding to date to $8.1 million.

At enterprise companies, LLM usage has surged alongside traditional ML use cases. At the heart of both these systems is private data. The most critical thing that ML teams do is take their raw data and transform it into valuable signals to feed into LLMs via prompts or ML models as inputs. Featureform believes there needs to be a unified framework to define, manage, and deploy these signals (or features). This creates a unified resource library that can be used by all ML/AI teams across an organization with built-in search & discovery, monitoring, orchestration, and governance. Featureform has seen to this be true with their existing customers in the ML space and has begun spearheading this approach in the LLM space.

MLOps is moving out of the hype phase and entering the actual productivity phase, says Featureform Founder and CEO Simba Khadder. On the backend of this, were seeing a huge wave of new use-cases that have been unlocked with LLMs. Data is at the core of these two systems, and in practice, the problems look almost identical. Featureforms frameworks will fundamentally change the way ML and AI teams work with data.

The rise of Retrieval Augmented Generation architecture, or RAG, has provided a way for data scientists to inject relevant signals and content from their data sets into their prompts as content to increase an LLMs accuracy and decrease likelihood of hallucination. These signals are analogous to traditional machine learning features that youd find in a feature store. The big difference is that, after being processed, they are stored in a vector database. By adding vector database support, Featureform becomes the hub where data scientists can define, manage, and deploy their features for both ML and LLM systems.

Featureforms feature store platform offers a distinct advantage in the market with its unique virtual architecture, says Gautam Krishnamurthi, Partner at GreatPoint Ventures. This virtual approach not only sets them apart from the competition, but also significantly lowers the cost of feature store implementation in the large and growing MLOps market. Coupled with their expert team, Featureform provides a best-in-class solution in the market for building out machine learning feature management. We are excited to support the Featureform team in their Seed round and beyond!

Featureform provides data scientists with a framework to turn their data into useful features for ML models and LLMs. By using Featureform, these teams:

To learn more visit https://featureform.com.

About Featureform

Featureform is the creator of the virtual feature store. Our mission is to streamline how data and model features are built and maintained in machine learning orgs. Our python framework and feature store does away with copy and pasting between scattered notebooks with names like Untitled18.ipynb, unifies feature pipelines between experimentation and production, deduplicates repeated features across teams, and eliminates ambiguously named tables like feature_table_v5. While we pride ourselves on our open-core model, we also offer a robust enterprise solution with governance, streaming, and more. We are proudly based out of San Francisco.

Source: Featureform

Continued here:
Featureform Raises $5.5M by Refining and Accelerating the Way Teams Work on AI and ML - Datanami

Read More..

Adapt or Extinct: Custom GPTs and the Evolution of Startups – Adapt or Extinct: Custom GPTs and the Evolution of … – InformationWeek

In the vast landscape of technological evolution, the recent unveiling of OpenAI's Custom GPTs can be likened to a meteoric event, shaking the very foundations of the startup ecosystem. Much like Steven Spielberg's Life on Our Planet, where the first episode is aptly titled The Rules of Life, the dynamics in the startup realm are undergoing a seismic shift. The question that looms large is: Did a mass extinction of startups just occur?

Spielberg's first rule, The best adapted will always win through, resonates deeply in the tech ecosystem. Large language models (LLMs) are now environmental, but Custom GPTs present a new set of rules. Companies that build deep value on top of LLMs will become the best adapted in this scenario - because they can provide insights that OpenAI cannot.

The second rule, Competition drives adaptation, takes center stage as startups face a new contender in the form of OpenAI's Custom GPTs. The ability to harness the power of AI is now a prerequisite for survival. Those who embrace and integrate this technology will find themselves on the winning side of the competition, echoing the age-old principles of evolution.

The third rule, Earth never remains stable for long, draws a parallel to the rapidly changing landscape of the startup industry. OpenAI's announcement sends ripples across the entrepreneurial world, causing both chaos and opportunity. The environment will continue to shift, but the economy - and business interest in products -- moves on.

Related:ChatGPT: Benefits, Drawbacks, and Ethical Considerations for IT Organizations

And while weve been focused on startups and the impact Custom GPTs will have on them, the bigger challenge might be for incumbents. Startups are nimble and many have built their solutions on AI as native underpinning components, giving them not only a broader arsenal but making them formidable players in technology. Incumbents, who tend to be slower to adopt and incorporate new technologies into their offerings, may find themselves akin to the dinosaurs if they do not find a way to adapt and catch up quickly. The acceleration of technological evolution is like the rules of life that have governed our planet for eons -- a perpetual cycle of competition and adaptation. It doesnt matter if a company was founded yesterday or twenty years ago -- the one that offers the best product and that delivers the most value will always win out.

If startups see this change in the ecosystem as an opportunity, instead of a threat, they are already starting on the path to success and survival. Custom GPTs should not be viewed as a replacement for a task -- they are tools that elevate the capabilities of those building upon them. And while this means that the simpler applications might not be as needed, it will empower more startups to build solutions that solve more complex problems. And there is an opening for that -- you can't white-label a Custom GPT or seamlessly integrate it into your website. Startups just need to reframe their mentality. Instead of focusing on what Custom GPTs can do, they should focus on what they cant do and adjust the path ahead accordingly.

Related:How Could AI Be a Tool for Workers?

And its important to note that after the meteor, comes the ice age. Meaning, it is unlikely that this will be the only move by OpenAI, or another company, that will disrupt the startup ecosystem. More innovations will come, and more startups will have to learn to pivot in response. Those who remain nimble, who stay on the cutting edge of innovation, and who deliver real results and value to their customers will continue to evade extinction.

The emergence of Custom GPTs is more than a technological breakthrough; it is a transformative moment in the business world. It signifies a call to evolve, to become AI natives, embedding AI deeply in business ethos and operations. As history has repeatedly shown, in the face of evolutionary change, those who adapt most effectively lead the way forward.

Related:Making the Most of Generative AI in Your Business: Practical Tips and Tricks

OpenAI's move is not a death knell for startups; instead, it is a catalyst that propels them into uncharted territories of innovation. The meteor may have hit, but it's the small, adaptable creatures that stand a chance at survival. The startups, with their agility and ability to respond, are not facing extinction but rather a rejuvenation as we enter a new era with new winners set to emerge.

Read this article:
Adapt or Extinct: Custom GPTs and the Evolution of Startups - Adapt or Extinct: Custom GPTs and the Evolution of ... - InformationWeek

Read More..

Apple joins the AI war with robust Machine Learning Framework | Philip Okoampah Kwaning – Medium

Is Apple planning to enter the AI race?

On December 6, 2023, Google shook the tech world with its grand AI ambitions, unveiling Gemini, its largest AI model from the DeepMind division. Not wanting to lag behind, Apple quietly entered the AI arena, introducing its own machine learning (ML) framework called MLX. While Googles Gemini grabbed headlines, Apples move suggests a subtle but significant step into the AI landscape.

The recently launched MLX is a machine learning framework designed for developers to efficiently build models on Apple Silicon. Paired with the MLXData deep learning model library, it hints at Apples potential venture into creating its own AI foundation models. Unlike the generative AI models released by Google and OpenAI, Apple remains discreet about its artificial intelligence endeavors, often referring to them as machine learning.

According to reports, MLX isnt for consumers but empowers Apples developers with a robust environment for training ML models. What sets it apart is Apples flexibility, allowing developers to choose their preferred coding language without constraints. Apple ML researcher Awni Hannum shared the release on GitHub, emphasizing its efficiency on Apple silicon devices.

In a move towards transparency, Apple has made MLX an open-source framework, inviting developers to explore and contribute. The framework seamlessly integrates with popular ML frameworks like PyTorch, ArrayFire, or Jax, making it accessible for a broader community.

Apple emphasizes that MLX aims to be user-friendly while ensuring efficient model training and deployment. The accompanying note suggests Apples intent to encourage researchers to enhance MLX, fostering quick exploration of new ideas in the ever-evolving field of machine learning.

The true impact of MLX will unfold over time as developers leverage this framework to create foundation models. Apples step into the AI domain signifies a strategic move, and only time will reveal the unique contributions and directions this release takes within the evolving landscape of artificial intelligence. Stay tuned for updates on the exciting developments spurred by Apples MLX framework.

Read the original here:
Apple joins the AI war with robust Machine Learning Framework | Philip Okoampah Kwaning - Medium

Read More..

Refining Non-Invasive Diagnosis of IPF: Development and Validation of a CT-Based Deep Learning Algorithm – Physician’s Weekly

The following is a summary of Development and validation of a CT-based deep learning algorithm to augment non-invasive diagnosis of idiopathic pulmonary fibrosis, published in the November 2023 issue of Pulmonology by Maddali, et al.

For a study, researchers sought to enhance the non-invasive diagnosis of idiopathic pulmonary fibrosis (IPF) by developing and validating a machine learning algorithm utilizing computed tomography (CT) scans exclusively. The aim was to improve the identification of the usual interstitial pneumonia (UIP) pattern and distinguish IPF from other interstitial lung diseases (ILD), reducing the need for invasive surgical biopsies.

A primary deep learning convolutional neural network (CNN) was employed, trained on a diverse multi-center dataset of over 2000 ILD cases with a reference standard of multidisciplinary discussion (MDD) consensus diagnosis. The algorithm was fine-tuned on a US-based multi-site cohort (n = 295) and externally validated with a separate dataset (n = 295) from European and South American sources.

In the tuning phase, the developed machine learning model demonstrated a commendable performance with an area under the receiver operating characteristic curve (AUC) of 0.87 (CI: 0.830.92) for distinguishing idiopathic pulmonary fibrosis (IPF) from other interstitial lung diseases (ILDs). The sensitivity and specificity reached 0.67 (0.570.76) and 0.90 (0.830.95). Notably, the model outperformed pre-recorded assessments conducted before multidisciplinary discussion (MDD) diagnosis, where sensitivity was only 0.31 (0.230.42), and specificity was 0.92 (0.870.95). The external test set validated the models robustness, yielding a c-statistic of 0.87 (0.830.91). Remarkably, the models performance consistency extended across diverse CT scanner manufacturers and various slice thicknesses.

The deep learning algorithm, relying solely on CT images, accurately identified IPF within ILD cases. The consistent results across diverse datasets and scanner variations suggested its potential as a valuable tool for non-invasive IPF diagnosis, offering improvements over traditional diagnostic approaches.

Source: resmedjournal.com/article/S0954-6111(23)00316-5/fulltext

See the original post:
Refining Non-Invasive Diagnosis of IPF: Development and Validation of a CT-Based Deep Learning Algorithm - Physician's Weekly

Read More..

Google Gemini Pro helping devs and organisations build AI – Technology Magazine

We are introducing a number of important new capabilities across our AI stack in support of Gemini, our most capable and general model yet, said Thomas Kurian, CEO of Google Cloud. It was built from the ground up to be multimodal, which means it can generalize and seamlessly understand, operate across, and combine different types of information, including text, code, audio, image, and video, in the same way humans see, hear, read, listen, and talk about many different types of information simultaneously.

Google has also announced it has updated Vertex AI with Imagen 2, its most advanced text-to-image diffusion technology from Google DeepMind to date. The company says Imagen 2 delivers significantly improved image quality and a host of features, including the ability to generate a wide variety of creative and realistic logos including emblems, lettermarks and abstract logos for business, brands and products, and can deliver improved results in areas where text-to-image tools often struggle, including rendering text in multiple languages.

Imagen 2 on Vertex AI boasts a variety of image generation features to help organizations create images that match their specific brand requirements with the same enterprise-grade reliability and governance customers are used to with Imagen.

Snap is using Imagen to help Snapchat+ subscribers to express their inner creativity. With the new AI Camera Mode, users can tap a button, type in a prompt or choose a pre-selected one, and generate a scene to share with family and friends or post to their story.

Imagen is the most scalable text-to-image model with the safety and image quality we need, said Josh Siegel, Senior Director of Product at Snap. One of the major benefits of Imagen for Snap is that we can focus on what we do best, which is the product design and the way that it looks and makes you feel. We know that when we work on new products like AI Camera Mode, we can really rely on the brand safety, scalability, and reliability that comes with Google Cloud.

Shutterstock has also emerged as a leading innovator bringing AI to creative production including being the first to launch an ethically-sourced AI image generator, now enhanced with Imagen on Vertex AI. With the Shutterstock AI image generator, users can turn simple text prompts into unique, striking visuals, allowing them to create at the speed of their imaginations. The Shutterstock website includes a searchable collection of more than 16,000 Imagen pictures, all available for licensing.

We exist to empower the world to tell their stories by bridging the gap between idea and execution. Variety is critical for the creative process, which is why we continue to integrate the latest and greatest technology into our image generator and editing featuresas long as it is built on responsibly sourced data, said Chris Loy, Director of AI Services, Shutterstock. The Imagen model on Vertex AI is an important addition to our AI image generator, and we're excited to see how it enables greater creative capabilities for our users as the model continues to evolve.

******

For more insights into the world of Technology - check out the latest edition of Technology Magazine and be sure to follow us on LinkedIn & Twitter.

Other magazines that may be of interest - AI Magazine | Cyber Magazine | Data Centre Magazine

Please also check out our upcoming event - Sustainability LIVE Net Zero on 6 and 7 March 2024.

******

BizClik is a global provider of B2B digital media platforms that covers executive communities for CEOs, CFOs, CMOs, sustainability leaders, procurement & supply chain leaders, technology & AI leaders, fintech leaders as well as covering industries such as manufacturing, mining, energy, EV, construction, healthcare and food.

Based in London, Dubai, and New York, Bizclik offers services such as content creation, advertising & sponsorship solutions, webinars & events.

See the article here:
Google Gemini Pro helping devs and organisations build AI - Technology Magazine

Read More..