Page 1,544«..1020..1,5431,5441,5451,546..1,5501,560..»

Malware and machine learning: A match made in hell – Help Net Security

Weve been developing machine learning-based cybersecurity systems for many years and began developing automation for analysis in our labs in 2005. These early automation projects have since evolved into full-blown machine-learning frameworks. Since then, weve been waiting for our enemies to make the same move, and after 18 years, the wait is over malware with artificial intelligence has arrived.

Defenders have been able to automate their work for some time, enabling excellent detection, analysis and reaction times hands-free at machine speed. This contrasts with attackers who have had to build and deploy their attacks manually, meaning that when they get blocked, they have to change things manually at much slower human speed.

The technology to run malware campaigns and automatically bypass new defenses is most definitely doable nowadays, but thus far, we havent seen anything of the kind. However, when it does happen, it will be noticeable as it would clearly signal that our enemies reaction speed has changed from human to machine-speed.

Deepfakes are probably the first thing that comes to mind when discussing AIs criminal or malicious use. Nowadays, its easy to create realistic images of fake people, and we see them frequently used in romance scams and other fraud cases. However, deep fakes of real people are something different altogether, and while abuse of deep fake images, voices and videos is, thus far, relatively small in scale, there is no doubt that this will get worse.

Large language models (LLMs) like GPT, LAMDA and LLaMA are not only able to create content in human languages, but in all programming languages, too. We have just seen the first example of a self-replicating piece of code that can use large language models to create endless variations of itself.

How do we know about this? Because the malwares author SPTH mailed it to me.

This individual is what we would call an old-school virus hobbyist, and they would appear to like writing viruses that break new ground. SPTH has also created a long list of malware over the years, such as the first DNA-infecting malware, Mycoplasma Mycoides SPTH-syn1.0. However, it should be stressed that SPTH only seems to do this because they can and dont seem to be interested in using the malware to cause damage or steal money.

SPTHs self-replicating code is called LLMorpher. SPTH recently wrote: Here we go beyond and show how to encode a self-replicating code entirely in the natural language. Rather than concrete code instructions, our virus consists of a list of well-defined sentences which are written in English. We then use OpenAIs GPT, one of the most powerful artificial intelligence systems publicly available. GPT can create different codes with the same behaviour, which is a new form of metamorphism.

This piece of code is able to infect programs written in the Python language. When executed, it searches the computer for .py files and copies its own functions into them. However, the functions are not copied directly; the functionality is described in English to GPT, which then creates the actual code that gets copied. This results in an infected Python file, which will keep replicating the malware to new files and the functions are reprogrammed every time by GPT something never been done before.

Simply writing malware is not illegal; using it to infect peoples systems or to cause damage is. So while SPTH doesnt appear to have done anything like that, this is still very problematic because third parties can misuse SPTHs research; LLMorpher can be easily downloaded from Github.

LLMorpher cant work without GPT. It doesnt have a copy, as GPT is unavailable for download. This means that OpenAI (GPTs creator) can simply block anyone using GPT for malicious purposes. Some similar models are downloadable (LLaMA, for example), so we will probably see those embedded into malware eventually.

Detecting malicious behavior is the best bet against malware that uses large language models, and this is best done by security products, which also use machine learning!

The only thing that can stop a bad AI is a good AI.

Continue reading here:
Malware and machine learning: A match made in hell - Help Net Security

Read More..

5 Key Learnings about AI and ChatGPT in the Enterprise – The New Stack

If 2022 was the year when AI broke through to become a society-changing technology, 2023 has been the year of AI breaking through in the enterprise. Put it this way: generative AI and large language models (LLMs) have become a familiar part of the lexicon for IT departments the world over. Your CIO is now more likely to namecheck ChatGPT than Kubernetes.

Already, its been a year of giant strides forward for AI in the enterprise, so here are five things weve learned so far.

In March, OpenAI announced an enterprise version of ChatGPT. But in this case, OpenAI was a fast follower not the first-to-market. Cohere, a Toronto-based company with close ties to Google, was already selling generative AI to businesses, as I discovered when I spoke to them that month.

Were working with developers in organizations, the AI/ML teams, to bring these capabilities into their organizations, Coheres CEO Martin Kon told us. He added that its approach is fundamentally different from OpenAIs. OpenAI wants you to bring your data to their models, exclusive to Azure. Cohere wants to bring our models to your data, in whatever environment you feel comfortable in.

So far, companies are mostly using generative AI to create semantic search engines for their own private data whether for internal use or for external customers. A related use case is knowledge management (KM). Imagine an employee being able to have a conversation with an AI that was trained on large language models based on company data.

One of the many new AI companies trying to achieve KM in chatbot form is Vectara. Its CEO Amr Awadallah told me that in five years, every single application whether that be on the consumer side or in the business/enterprise side will be re-architected in a way that is a lot more human in nature, in terms of how we express what we are trying to achieve and what were trying to do.

Last month, Google Cloud and Googles Workspace division announced a raft of new AI functionality. These included Generative AI App Builder, which allows organizations to build their own AI-powered chat interfaces and digital assistants, and new generative AI features in Google Workspace.

Google has coined a typically awkward term for applications powered by generative AI: gen apps. It claims that gen apps will become a third major internet application category, after web apps and mobile apps.

Im willing to wager that gen apps as a term will never take off but, regardless, Googles AI-powered tools will be well used within enterprise companies.

Likewise, Microsoft has been releasing new AI tools, such as Semantic Kernel (SK), described as an open-source project helping developers integrate cutting-edge AI models quickly and easily into their apps. SK is in many ways a classic Microsoft low-code tool, only it focuses on helping users do prompts for AI chatbots.

If you scan Stanfords HELM website, which measures LLMs in a variety of ways, youll see that these models vary greatly in size. There are, simply put, tradeoffs between model size and the speed it can work at.

OpenAI has several models, ranging in size from 1.3 billion parameters (its Babbage model) to 175B parameters (its DaVinci model).

Cohere goes as far as differentiating its model sizes as if they were Starbucks cups: small, medium, large, and extra large.

List of Coheres models in Stanford HELM directory

The primary models of OpenAI

However, Stanford also measures accuracy and in these statistics, size doesnt seem to matter much:

Stanford HELM tests for accuracy of ML models

In this new era of generative AI, frameworks like Ray are now just as important as Kubernetes in creating modern applications at scale. An open source platform called Ray offers a distributed machine-learning framework. Its being used by both OpenAI and Cohere to help train their models. Its also used in other highly-scaled products, such as Uber.

The company behind Ray is Anyscale, whose CEO Robert Nishihara told me earlier this year is very developer-centric. All of Rays functionality is designed to be easy to use for developers, he said. This differs from the Kubernetes user experience for devs, he noted, which is notoriously difficult. Ray was designed for Python developers, which is the main programming language being used in AI systems.

Anyscale co-founder Ion Stoica told me that Ray is like an extension of Python and like Python there are a set of Ray libraries that are targeted towards different use cases. The awkwardly named RLlib is for reinforcement learning, but there are similar libraries for training, serving, data pre-processing, and more.

Just as cloud computing ushered in a raft of big data solutions, generative AI is a catalyst for a new wave of data intelligence companies.

I recently spoke to Aaron Kalb, a co-founder of Alation, which styles itself as a data intelligence platform and promotes a concept it calls the data catalog. This combines machine learning with human curation to create a custom store of data for enterprise companies.

Kalb noted that both AI and the common enterprise acronym of BI (business intelligence) are garbage in, garbage out. He said that data intelligence is a layer that precedes AI and BI, that makes sure you can find, understand and trust the right data to put into your AI and BI.

In this context, he said, taking something like ChatGPT from the public internet and bringing it into the enterprise is very risky. He thinks that data needs to be, well, more intelligent before it is used by AI systems within an enterprise. Also, he doesnt think that the internet scale of ChatGPT and similar systems is needed in the enterprise. Every organization has its own terminology, he explained that could be industry terms, or things that are very specific to that company.

Its hard to believe that its been just a year since generative AI burst onto the scene. It all started with OpenAIs DALL-E 2, which was announced last April and launched as a private beta in July. DALL-E 2 is an image generation service powered by deep learning models, and it was a significant step forward for the industry. Also, last July, a company called Midjourney released its eponymous text-to-image generator.

The AI hype really ramped up in August, though, with the release of Stable Diffusion another deep-learning text-to-image generator. Unlike DALL-E 2 and Midjourney, Stable Diffusion had a permissive licensing structure. This was around the time people began paying more attention to the models behind all these services the LLMs. DALL-E 2 used a version of GPT-3, OpenAIs main LLM at the time.

But, in hindsight, we can see that the image generators were just the appetizer. At the end of November, OpenAI launched ChatGPT, a chatbot built on top of GPT-3.5. This was the catalyst for generative AI to enter the enterprise, and it has quickly taken over IT departments since then.

View post:
5 Key Learnings about AI and ChatGPT in the Enterprise - The New Stack

Read More..

The week in AI: The pause request heard round the world – TechCrunch

Image Credits: Andriy Onufriyenko / Getty Images

Keeping up with an industry as fast-moving as AI is a tall order. So until an AI can do it for you, heres a handy roundup of the last weeks stories in the world of machine learning, along with notable research and experiments we didnt cover on their own.

In one of the more surprising stories of the past week, Italys data protection authority (DPA) blocked OpenAIs viral AI-powered chatbot, ChatGPT, citing concerns that the tool breaches the European Unions General Data Protection Regulation. The DPA is reportedly opening an investigation into whether OpenAI unlawfully processed peoples data, as well as over the lack of any system to prevent minors from accessing the tech.

Its unclear what the outcome might be; OpenAI has 20 days to respond to the order. But the DPAs move could have significant implications for companies deploying machine learning models not just in Italy, but anywhere within the European Union.

As Natasha notes in her piece about the news, many of OpenAIs models were trained on data scraped from the internet, including social networks like Twitter and Reddit. Assuming the same is true of ChatGPT, because the company doesnt appear to have informed people whose data it has repurposed to train the AI, it might well be running afoul of GDPR across the bloc.

GDPR is but one of the many potential legal hurdles that AI, particularly generative AI (e.g. text- and art-generating AI like ChatGPT), faces. Its becoming clearer with each mounting challenge that itll take time for the dust to settle. But thats not scaring away VCs, who continue to pour capital into the tech like theres no tomorrow.

Will those prove to be wise investments, or liabilities? Its tough to say at present. Rest assured, though, that well report on whatever happens.

Here are the other AI headlines of note from the past few days:

At AI enabler Nvidia, BioNeMo is an example of their new strategy, where the advance is not so much that its new, but that its increasingly easy for companies to access. The new version of this biotech platform adds a shiny web UI and improved fine-tuning of a bunch of models.

A growing portion of pipelines are dealing with heaps of data, amounts weve never seen before, hundreds of millions of sequences we have to feed into these models, said Amgens Peter Grandsard, who is leading a research division using AI tech. We are trying to obtain operational efficiency in research as much as we are in manufacturing. With the acceleration that tech like Nvidias provides, what you could have done last year for one project, now you can do five or 10 using the same investment in tech.

This book excerpt by Meredith Broussard over at Wired is worth reading. She was curious about an AI model that had been used in her cancer diagnosis (shes OK) and found it incredibly fiddly and frustrating to try to take ownership of and understand that data and process. Medical AI processes clearly need to consider the patient more.

Actually nefarious AI applications make for new risks, for instance attempting to influence discourse. Weve seen what GPT-4 is capable of, but it was an open question whether such a model could create effective persuasive text in a political context. This Stanford study suggests so: When people were exposed to essays arguing a case in issues like gun control and carbon taxes, AI-generated messages were at least as persuasive as human-generated messages across all topics. These messages were also perceived as more logical and factual. Will AI-generated text change anyones mind? Hard to say, but it seems very likely that people will increasingly put it to use for this kind of agenda.

Machine learning has been put to use by another group at Stanford to better simulate the brain as in, the tissue of the organ itself. The brain is not just complex and heterogeneous, but much like Jell-O, which makes both testing and modeling physical effects on the brain very challenging, explained professor Ellen Kuhl in a news release. Their new model picks and chooses between thousands of brain modeling methods, mixing and matching to identify the best way to interpret or project from the given data. It doesnt reinvent brain damage modeling, but should make any study of it faster and more effective.

Out in the natural world,a new Fraunhofer approach to seismic imaging applies ML to an existing data pipeline that handles terabytes of output from hydrophones and airguns. Ordinarily this data would have to be simplified or abstracted, losing some of its precision in the process, but the new ML-powered process allows analysis of the unabridged dataset.

Interestingly, the researchers note that this would ordinarily be a boon to oil and gas companies looking for deposits, but with the move away from fossil fuels, it can be put to more climate-friendly purposes like identifying potential CO2 sequestration sites or potentially damaging gas buildups.

Monitoring forests is another important task for climate and conservation research, and measuring tree size is part of it. But this task involves manually checking trees one by one. A team at Cambridge built an ML model that uses a smartphone lidar sensor to estimate trunk diameter, having trained it on a bunch of manual measurements. Just point the phone at the trees around you and boom. The system is more than four times faster, yet accurate beyond their expectations, said lead author of the study, Amelia Holcomb: I was surprised the app works as well as it does. Sometimes I like to challenge it with a particularly crowded bit of forest, or a particularly oddly shaped tree, and I think theres no way it will get it right, but it does.

Because its fast and requires no special training, the team hopes it can be released widely as a way to collect data for tree surveys, or to make existing efforts faster and easier. Android only for now.

Lastly, enjoy this interesting investigation and experiment by Eigil zu Tage-Ravn of seeing what a generative art model makes of the famous painting in the Spouter-Inn described in chapter 3 of Moby-Dick.

Original post:
The week in AI: The pause request heard round the world - TechCrunch

Read More..

What Is Deep Learning AI & How Does It Work Forbes Advisor INDIA – Forbes

There is a lot of buzz around artificial intelligence and its different algorithms. We all are quite aware that machines along with specific computer algorithms can do wonders in our homes, offices or at workplaces. With the advancement of technology, one must know the main reasons behind several hi-tech inventions and innovations, is the new concept of deep learning.

Ever wondered how Netflix or Amazon showing recommendations based on your preferences, or how Siri is able to respond to each and every request. It means, deep learning algorithms are at work.

Right now, it may leave you a little perplexed, but heres a simple guide on deep learning, how it works and how it is deeply associated with the AI world. Forbes Advisor India has broken down this particular branch of AI in most simple terms.

Deep learning runs many artificial intelligence (AI) applications and services. It helps in adding intelligence and improving automation to the existing AI enabled products. DL is that part of AI which helps in performing analytical and physical tasks without any sort of human intervention.

In short, deep learning is a complex technique of machine learning, which instructs computers to learn or respond as to what naturally comes to humans. So, whether it is driverless cars, hands-free speakers, voice recognition in phones, tablets, TV or watches, deep learning is a major force behind all these breakthrough innovations.

In the concept of deep learning, the computer learns to perform on the basis of direct data feed such as image, text or sound. Such models are capable of achieving super accurate results and sometimes much better and more efficiently than human beings. Models based on deep learning uses a large set of data which requires high computation power and responds accurately via using a neural network which contains multiple layers like that of the humans brain.

In nutshell, deep learning sits inside of machine learning, which sits inside of artificial intelligence.

Artificial Intelligence: The development of a computer system which is able to perform all the given tasks at par with human intelligence.

Machine Learning: It is a subset of AI which contains statistical algorithms which enable machines to improve the tasks with experience. As opposed to deep learning, machine learning models need human intervention to improve accuracy.

Deep Learning: It is a kind of machine learning which has a human brain-like structure. It works on the basis of logical assembly of algorithms which are known as neural networks.

In the world of efficiency and accuracy, deep learning has made a notable and dominant position than ever before. Deep learning applications are used in various industries from healthcare, automated driving, medical devices, aerospace and defense, electronics and industrial automation.

Deep learning is extensively used in automated hearing, speech recognition, language translation, digital assistance, etc.

To give such accurate results, DL requires a large amount of labeled data and high computation power. High-performance GPUs have a perfect and ideal architecture which has been proved efficient for deep learning to perform. When combined with clusters or cloud computing, this technology enables teams to reduce training time for a deep learning network from weeks to hours or may be lesser than this.

For example, driverless car development requires billions of images and hours and hours of video, which help deep learning to automatically detect objects such as pedestrians, stop signs and traffic lights.

Mostly, all the deep learning-based models use architectures related to neural networks. The term deep is associated with this technique as it refers to the multiple number of hidden layers which are present in a neural network.

Deep learning models have been given proper training by using a large subset of labeled data and neural network architectures, which in turn helps these models to learn directly from that data without the need of any human intervention.

Neural networks are generally organized in multiple layers consisting of a different set of interconnected nodes. It is to be noted that these networks have the ability to have tens or hundreds of hidden layers.

Most deep learning features use the transfer learning approach, a procedure which involves fine-tuning a pretrained model. However, the relevant features are not pre trained as they are learned while the whole network trains on a collection of images. This feature includes automated extraction which makes deep learning models very accurate.

The key advantage of deep learning models is that they continue to improve as the size of the data upsurges. The more practice deep-learning algorithms get, the better they become.

Deep learning algorithms are very complex in their nature. There are different types of neural networks which helps in addressing the specific problems or datasets, such as:

Convolutional Neural Network (CNN): This kind of neural network has a certain degree of complexity which is seen in human brains. Neural network is not just made up of one layer but it consists of various layers which are also known as additional convolutional layers or pooling layers. It is to be noted that each layer plays a very crucial role in grasping the data and thus reaching a final conclusion which means identifying bigger portions of the image.

The first layer focuses on simple and basic features, such as colors and as the data progresses the layers of the network start to recognize much bigger elements of the object until it finally identifies the final object.

Recurrent Neural Network (RNN): Recurrent neural networks use consecutive data or time series data to resolve common problems seen in language translation and speech recognition which are used in various applications such as Siri, voice search, and google translate.

Similar to convolutional neural networks (CNNs), the RNN uses training data to learn. They distinguished the fed data by their memory as they take cues from prior inputs, which in turn influences the current input and output.

As we have already learnt, deep learning requires a massive amount of computation power. For this, high performance graphical processing units or (GPUs) are considered as ideal as they are able to handle and process quite a large volume of calculations and patterns with abundant memory available. It is to be noted that managing multiple such GPUs need high requirement of of internal resources which can be incredibly costly

Deep learning plays an important role in AI predictive modeling. It collects massive amounts of data and analyzes it to create multiple predictive models by understanding various patterns and trends within the data. Deep learning models make it very fast and easy to construct large amounts of data and form them into meaningful information. It is widely used in multiple industries, including automatic driving and medical devices. It is just a matter of time as this technology continues to mature.

Read more:
What Is Deep Learning AI & How Does It Work Forbes Advisor INDIA - Forbes

Read More..

New method uses machine learning for more robust fluid dynamics simulations – Imperial College London

A new method that uses advanced machine learning techniques can improve the accuracy of predictions from computational fluid dynamics simulations.

Machine learning experts from Imperials Data Science Institute as part of the INHALE project, have developed a new workflow applicable to any research or engineering field that uses computational fluid dynamics, from designing aircraft and cars, to predicting weather patterns and air pollution levels.

There is a growing interest in the use of machine learning techniques to improve the accuracy and efficiency of computational fluid dynamics a field that involves using numerical methods to solve complex equations describing the behaviour of fluids such as air or water.

By applying their new framework on two different case studies, the researchers including Research Associate Dr Csar Quilodrn Casas and Lecturer in machine learning Dr Rossella Arcucci were able to demonstrate the effectiveness of their proposed model using real-world scenarios involving air pollution flows.

The study, published in Physica A: Statistical Mechanics and its Applications uses deep learning - a subfield of machine learning, and a technique used in deep learning called adversarial training which helps to make the models more trustworthy.

Traditional computational fluid dynamics simulations can be computationally expensive and time consuming, which limits their use in real-world applications.

Surrogate models are often used to address these challenges, providing a simplified version of a computationally expensive model whilst still producing accurate predictions or simulations. However, it is difficult to create surrogate models that accurately capture the behaviour of fluids, especially in complex scenarios such as monitoring air pollution flows.

Machine learning techniques offer a promising solution to this problem by creating surrogate models that can accurately predict the behaviour of fluids with much less computational cost.

In the method, the DSI team use a machine learning technique called adversarial training to address the aforementioned challenges, creating surrogate models that are more accurate and efficient than those produced through traditional methods.

Adversarial training is a technique used in deep learning to make computer models more robust against attacks that deliberately try and deceive them. This involves training the model on both the original data and data that has been modified to stimulate an attack.

By exposing the model to these adversarial examples, it can learn to recognise and ignore them, making the model more resistant to future attacks.

Through using adversarial training in their model, the researchers were able to provide surrogate models that are more accurate and efficient, even with limited data available for training.

According to Lead Author Dr Quilodrn-Casas: The new workflow we have created has the ability to help and assist engineers and modellers towards creating cheap and accurate model surrogates of expensive Computer Fluid Dynamics simulations, not necessarily just for air pollution

This study demonstrates the effectiveness of adversarial training for creating surrogate models that accurately capture the behaviour of fluids in complex scenarios such as urban air pollution flows.

The teams findings could also be useful to engineers and designers who work with fluid systems, such as aircrafts and cars, and who are interested in optimising their designs using computational models.

Following on from this study, Dr Quilodrn-Casas will be using the new workflow upcoming research into wave energy flows and beyond this, the method could be applied to other domains aside from computational fluid dynamics such as weather forecasting.

-

A data-driven adversarial machine learning for 3D surrogates of unstructured computational fluid dynamic simulations by Quilodrn-Casas and Arcucci, published on 21 February 2023 in Physica A: Statistical Mechanics and its Applications.

More:
New method uses machine learning for more robust fluid dynamics simulations - Imperial College London

Read More..

Exploring the Possibilities of IoT-Enabled Quantum Machine Learning – CIOReview

With quantum machine learning, the internet of things can become even more powerful, enabling people to create more efficient and safer systems.

FREMONT, CA: The Internet of Things (IoT) is altering how people interact with their surrounding environment. From intelligent homes to autonomous vehicles, the possibilities are limitless. Researchers are investigating the possibility of merging IoT with quantum machine learning (QML) to create even more powerful and efficient systems.

QML is an artificial intelligence (AI) that processes data using quantum computing. It offers the ability to provide quicker and more precise decision-making than conventional AI. Researchers hope to create a potent new data analysis and prediction tool by merging it with the IoT.

QML and IoT could be combined to create smarter, more efficient systems for various applications. For instance, it might optimize city traffic flow by forecasting traffic patterns and modifying traffic light timing accordingly. It could also be utilized to optimize building energy consumption and monitor and predict disease spread

IoT facilitates the huge potential of QML enabled by IoT. It could transform how people interact with the environment around them and create new opportunities for data analysis and forecasting. As researchers continue to investigate the possibilities, it is evident that this technology can alter the way of life.

Using the IoT to Advance QML

The IoT is altering how people interact with their surrounding environment. IoT technology's potential applications appear limitless, from intelligent homes to self-driving vehicles. Now, scientists are investigating how IoT can transform QML.

QML is a fast-developing research topic that blends quantum computing capabilities with machine learning methods. QML can enable robots to learn more effectively and precisely than ever before by harnessing the potential of quantum computing.

The IoT is ideally suited to supporting QML applications. IoT devices can collect and communicate vast quantities of data, which can be utilized to train and optimize machine learning algorithms. In addition, IoT devices can be used to monitor and control the environment in which QML algorithms are deployed, ensuring that they operate under optimal conditions.

Also, researchers are investigating how IoT devices might be leveraged to enhance the security of QML applications. IoT devices can identify and prevent harmful attacks on QML systems by harnessing the power of distributed networks. IoT devices can also be used to monitor the performance of QML algorithms, enabling the immediate identification and resolution of any problems.

The potential uses of the IoT for QML are vast, and researchers are just beginning to investigate them. By leveraging the power of the IoT, researchers are paving the way for a new era of QML that might transform how people interact with the world.

Originally posted here:
Exploring the Possibilities of IoT-Enabled Quantum Machine Learning - CIOReview

Read More..

US Army pursue AI and machine learning integration – Army Technology

The US Department of Defense (DoD) has been contracting big data analytics firm Palantir Technologies for research and development (R&D) services into the integration of artificial intelligence (AI) and machine learning (ML).

Their work has been ongoing since 28 September 2022, with recurring funds every few months, and is expected to be completed by 28 September 2023. Palantir was awarded $42m in their intital 2022 contract, with a further $10m on 14 December 2022, and the latest boost comprised $60m announced on 31 March 2023.

In their thematic intelligence report on Artificial Intelligence (2023), GlobalData estimates the total AI market will be worth $383.3bn in 2030, implying a 21% compound annual growth rate between 2022 and 2030. In the coming decade, the country that emerges on top in AI will lead the Fourth Industrial Revolution according to GlobalData. This area of the tech war between the US and China, in which the former has placed export controls of semi-conductor components on the latter.

ML is an application of AI that has the most practical benefits, as it allows computer systems to learn and improve from data without explicit programming.

The report asserts that the rapid growth in the volume of data is prompting demand for computing resources to analyse that data and this is a growing issue in the defence industry where command, control, communications, computers, intelligence, surveillance and reconnaissance (C4ISR) must process enormous amounts of data. AI and ML provide the means by which useful data can be acted upon or used without any substantial delay. In the battlespace environment, this provides the military with a great competitive edge.

The contracting body for the recently awarded funds is the US Army Contracting Command, Aberdeen Proving Ground, Maryland. It is this body that is in charge of the DoDs Distributed Common Ground System Army (DCGS-A).

This system is the US Armys primary system for intelligence, surveillance and reconnaissance (ISR) tasking of sensors, posting of data processing information and using intelligence information about threat, weather and terrain.

As the processing hub of a huge deluge of data the DCGS-A is the system that would benefit the most from the automated processing services that AI and ML can provide. In all likelihood Palantir will be researching AI application integration for the Army.

See the original post:
US Army pursue AI and machine learning integration - Army Technology

Read More..

Why Sophistication Will Win Out In The Machine Learning Ops Sector – Unite.AI

Theres no question that machine learning operations (MLOps) is a burgeoning sector. The market is projected to reach $700 million by 2025 almost four times what it was in 2020.

Still, while technically sound and powerful, these solutions havent generated the expected revenue, which has raised concerns about future growth.

I can understand the pessimism surrounding the space, as I spent the first 20 years of my career effectively building internal MLOps tools at an esteemed investment management firm. More recently, Ive invested in MLOps startups, but they have been slow to achieve the level of revenue that I would have expected. Based on both my positive and negative experiences with MLOps, I understand why these startups have struggled and why they are now poised for growth.

MLOps tools are critical to companies deploying data-driven models and algorithms. If you develop software, you need tools that allow you to diagnose and anticipate problems with software that could cause you to lose meaningful revenue due to its failure. The same is true for companies that build data-driven solutions. If you dont have adequate MLOps tools for evaluating models, monitoring data, tracking drift in model parameters and performance, and tracking the predicted vs. actual performance of models, then you probably shouldnt be using models in production-critical tasks.

However, companies deploying ML-driven solutions without deep knowledge and experience dont recognize the need for the more sophisticated tools and dont understand the value of the low-level technical integration. They are more comfortable with tools operating on externalities, even if they are less effective, since they are less intrusive and represent a lower adoption cost and risk if the tools dont work out.

On the contrary, companies with ML teams who possess deeper knowledge and experience believe they can build these tools in-house and dont want to adopt third-party solutions. Additionally, the problems that result from MLOps tools shortcomings arent always easy to identify or diagnoseappearing as modeling versus operations failures. The outcome is that companies deploying ML-based solutions, whether technically sophisticated or inexperienced, have been slow to adopt.

But things are starting to change. Companies are now recognizing the value of sophisticated, deeply integrated MLOps tools. Either they have experienced problems resulting from not having these tools or they have seen competitors suffering from their absence in many high-profile failures, and are now being forced to learn about the more complex MLOps solutions.

Those MLOps companies that have survived the revenue winter so far should see a thawing of the market and a growth in sales opportunities.

Companies selling superficial solutions will start losing business to more integrated solutions that are harder to understand and adopt, but provide more monitoring, debugging, and remediation services for their customers. MLOps software developers should keep the faith that building powerful software that solves problems in a deeper and more thorough way will win out in the long run over simple solutions that give immediate payoffs but dont solve the full breadth of problems their customers are facing.

Read more:
Why Sophistication Will Win Out In The Machine Learning Ops Sector - Unite.AI

Read More..

Predictive machine learning approaches for the microstructural … – Nature.com

Multiple-slip crystal plasticity dislocation-density

A dislocation-density crystalline plasticity approach was used in conjunction with FEM to generate the database used in this study. The crystalline plasticity approach was developed by Zikry and Kao20 and Shanthraj and Zikry21 and uses a set of partial differential equations to describe the dislocation evolution within a unit area called dislocation density. Separate equations are used for mobile and immobile dislocation densities, ({rho }_{m}) and ({rho }_{im}), and a set of nondimensional coefficients are used to describe the sourcing, trapping, annihilation, immobilization, and recovery of dislocations21. The dislocation density evolution equations are

$$frac{d{rho }_{m}^{alpha }}{dt}=left|{dot{gamma }}^{alpha }right|left(frac{{g}_{sour}^{alpha }}{{b}^{2}}left(frac{{rho }_{im}^{alpha }}{{rho }_{m}^{alpha }}right)-{g}_{mnter-}^{alpha }-{rho }_{m}^{alpha }-frac{{g}_{immob-}^{alpha }}{b}sqrt{{rho }_{im}^{alpha }}right),$$

(1)

$$frac{d{rho }_{im}^{alpha }}{dt}=left|{dot{gamma }}^{alpha }right|left({g}_{mnter+}^{alpha }{rho }_{m}^{alpha }+frac{{g}_{immob+}^{alpha }}{b}sqrt{{rho }_{im}^{alpha }}-{g}_{recov}{rho }_{im}^{alpha }right),$$

(2)

where gsour is the coefficient pertaining to an increase in the mobile dislocation density due to dislocation sources, gmnter is the coefficient related to the trapping of mobile dislocations due to forest intersections, cross slip around obstacles, or dislocation interactions, grecov is a coefficient related to the rearrangement and annihilation of immobile dislocations, and gimmob is related to the immobilization of mobile dislocations. These coefficients, which have been nondimensionalized, are summarized in Table 1, where ({f}_{0}), and (varphi) are geometric parameters. H0 is the reference activation enthalpy, lc is the mean free path of a gliding dislocation, b is the magnitude of the Burgers vector, and s is the saturation density. It should be noted that these coefficients are functions of the immobile and mobile densities, and hence are updated as a function of the deformation mode. Shear slip rate, (dot{gamma }), is a measure of the accumulated plastic strain on a material that is related to the mobile dislocation activity in a material as

$${dot{gamma }}^{(alpha )}={rho }_{m}^{(alpha )}{b}^{left(alpha right)}{v}^{left(alpha right)},$$

(3)

where ({v}^{(alpha )}) is the average velocity of mobile dislocations on slip system (alpha).

The orientation relationships (ORs) between the (delta) hydrides examined in this work and the surrounding matrix were developed in Mohamed and Zikry22 as

$${left(0001right)}_{hcp}// {left(111right)}_{fcc} and {left[11overline{2 }1right]}_{hcp}//{left(overline{1 }10right)}_{fcc},$$

(4)

and represent the plane relationships between the h.c.p. material and the f.c.c. material.

Each simulation was developed according to its own material fingerprint and consisted of a plane strain model with displacement control for the FE model. The average mesh size was 60,000 elements and consisted of 49 zirconium h.c.p. alloy grains and approximately 50 f.c.c. hydrides, for conditions including hydride spacing and hydride length based on the fingerprint. Strain was applied uniaxially, and grain orientation with respect to the loading axis was defined using the value of the material fingerprint parameter. The strain rate was constant at 10 ({s}^{-1}). Grain orientations were defined as the angle that the zirconium alloys [0 0 1 0] axis forms with respect to the loading axis, which is uniaxial at the [0 0 1 0] global direction. Changes to this parameter rotated the grain with respect to the [0 1 0 0] normal axis. Mohamed and Zikry23 validated the material properties used in the simulations, which are presented in Table 2.

To characterize the solution space of all possible material fingerprints, a total of 210 simulations were simulated using FEM. The material fingerprints for each simulation consisted of five material parameters and were chosen according to uniform distributions bounded by the values in Table 3. The parameter values were chosen from a grid of 4 equally spaced values from within these bounds. In addition to modifying grain orientation according to the range within these bounds, grain to grain misorientation was randomized at a maximum of 10. These parameters influence dislocation activity and fracture, and the specific values used correspond to experimental values4,11,23,24. To sample from the solution space, the trajectory method used in the Elementary Effects method (implemented in SAlib) was used25,26.

To avoid numerical issues with the model input parameters being at different physical scales, the parameters were processed by centering the mean of each feature around zero and scaling to unit variance using the StandardScaler function in Scikit-Learn (Version 0.23.2)27. The critical fracture stress values were also scaled by 100MPa.

When training the models, the data was randomly split into an 85% training set and a 15% validation set. The training set was used to train model hyperparameters. The hyperparameters were randomly chosen for training within predefined ranges chosen for each model type. Typically, 50,000 iterations within the hyperparameter space were used along with a fivefold cross validation technique to reduce overfitting to the data. After the best possible model hyperparameters were found, the model was tested on the validation set which comprised 15% of the original data, and which had not been used to train the model. The goodness of fit, for this model, is the measurement of the models performance in predicting this validation set.

A linear regression using OLS was used to provide a benchmark for the other modeling methods explored in this work. Because the simulations are non-linear, there was only a small likelihood of attaining a high level of accuracy with this class of models. However, linear regression provides the most interpretable model output of any of the other modeling systems. Scikit-Learns LinearRegression function was used to produce these models27.

The random forest regression model was used to generate a model. It is comprised of an ensemble of decision trees whose outputs are averaged. The result is a general model that provides accurate predictions in high dimensional spaces28. Decision tree-based methods are also helpful because their output can be interrogated, though it may be cumbersome to do so for an ensemble of decision trees. 100 estimators were used for each regression model, corresponding to 100 decision trees, which would make this kind of interpretation difficult. Other methods exist to interpret the output of a random forest regressor, and they are implemented here. The importance of each input parameter can also be determined using methods such as recursive feature elimination. Scikit-Learns RandomForestRegressor27 was used.

A multilayer perceptron, or neural network, was also fitted to the data. While neural networks are the least interpretable method presented in this study, they have also been shown to be powerful estimators for highly dimensional data. The models presented here were comprised of 3 hidden layers with 5 neurons each. Scikit-Learns MLPRegressor function was used to train and test these estimators27.

Gaussian Process Regression (GPR) was chosen as a model type because of its built-in measure of uncertainty. A combination kernel comprised of a Matern kernel and an Exponential Sine Squared kernel were used for training. The sinusoidal attribute of this kernel was a result of the prior understanding that material properties tend to follow a sinusoidal path as a non-isotropic material is rotated. The Matern kernel additionally allowed the model to effectively capture discontinuities in the solution space. The GaussianProcessRegressor function within the SciKit-Learn package was used to train the models, and a cross validated randomized search was used to find length scale, the Matern (nu) parameter, and the periodicity parameter for the exponential sine squared kernel27.

The purpose of this study is to obtain ROMs that describe the fracture stress state of a material given its material fingerprint and strain level. This is performed by predicting the (mu) value of a Gumbel distribution trained to the 95th percentile of each data set. These (mu) values are normalized by the fracture stress to provide physically based insights. These models can provide critical microstructural fracture predictions without FEM models or experimental measurements, and is a representation of incipient fracture within the material. The fracture critical stress information predicted from these models can then be used in conjunction with other computational and experimental methods to determine the likelihood of failure. These predictions are the link between the material fingerprint and the fracture probability for that material at a certain strain level.

Read more:
Predictive machine learning approaches for the microstructural ... - Nature.com

Read More..

How AI, machine learning and ChatGPT are changing the legal system – Bizcommunity.com

The field of technology law has seen significant growth over the past few years in South Africa. With the rapid pace of technological advancements and the increasing reliance on technology in various industries, the legal system must keep up with these changes. In this article, I will explore the future of tech law in South Africa.

One of the areas where technology law is likely to see development in South Africa is the regulation of data privacy. The Protection of Personal Information Act (PoPIA) protects personal information and regulates the processing of personal data. However, with the rise of big data and the increasing use of technology in various industries, the legal framework surrounding data privacy will likely evolve in the coming years. This may include changes to PoPIA itself, as well as new legislation and case law that addresses emerging issues in data protection. These issues include, but are not limited to

Another area where tech law will likely see development is regulating artificial intelligence (AI) and machine learning. As AI and machine learning become more widespread, concerns exist about their potential impact on the following areas:

In South Africa, no specific legislation currently governs the use of AI and machine learning. Therefore, to address these concerns, it is important for South African organisations and policymakers to prioritise privacy, security, and ethical considerations when developing and implementing AI and machine learning systems. This may involve developing robust data protection policies, ensuring adequate cybersecurity measures are in place, and promoting transparency and accountability in AI and machine learning systems.

The role of ChatGPT in the future of legal research and analysis cannot be overstated. As a large language model, ChatGPT has the potential to revolutionise the way legal research is conducted in South Africa. With the increasing volume of case law and legislation, it can be challenging for legal practitioners to keep up with developments in the field. ChatGPT can quickly and accurately analyse large volumes of legal texts, allowing legal practitioners to identify relevant case law and legislation. Additionally, ChatGPT can assist in legal writing by providing suggestions for legal arguments based on its analysis of previous case law and legislation.

In conclusion, the field of technology law in South Africa is likely to see significant growth and development in the coming years. With the rapid pace of technological advancements and the increasing reliance on technology in various industries, the legal system must keep up with these changes. The regulation of data privacy and AI and machine learning are two areas where tech law is likely to see development. Case law and legislation will be crucial in shaping these technologies' legal frameworks. ChatGPT also has the potential to revolutionise legal research and analysis in South Africa and will undoubtedly play a significant role in the future of tech law in the country.

Read the original here:
How AI, machine learning and ChatGPT are changing the legal system - Bizcommunity.com

Read More..