Page 3,419«..1020..3,4183,4193,4203,421..3,4303,440..»

Bitcoin ‘maximalists’ accused of ‘shilling’ an SEC-cleared token – Cointelegraph

On August 24, INX launched a tokenized initial public offering or IPO which was cleared by the SEC. The company describes its mission as bringing "regulated digital asset opportunities to institutions and retail investors".

The following day, several crypto influencers tweeted what appeared to be promotional statements about the company and its token. Stefan Jespers, known as WhalePanda on Twitter, compared the INX token to Binance's BNB:

Source: Twitter.

A similar sentiment was voiced byJameson Lopp, the CTO of Casa and a self-proclaimed cypherpunk:

Both Jespers and Lopp are considered Bitcoin maximalists a group that some categorize as holding negative attitudes towards altcoins and token offerings. Their statements left many feeling befuddled, with some surmising that the Twitter accounts in question may have been compromised. In the end, there was a more prosaic explanation for this unusual behavior, however. According to a tweet by CobraBitcoin, the long-time custodian of Bitcoin.org's website, the individuals in question had received INX options at $0.01 per token. He alleged that this would allow them to make a 90x profit during the IPO:

Source: Twitter.

Other notable members of the Bitcoin maximalist camp appeared as company advisors as well, including Alena Vranova, the founder of SatoshiLabs and Samson Mow, the chief strategy officer at Blockstream. All parties, with the exception of Mow, are listed on the company's website. Neither Vranova or Mow have tweeted about the exchange or its public offering.

Company advisors often receive various stock options as payment for their support. However, some may find it hypocritical that the same people who have accused others of selling "snake oil" are now promoting a token offering without offering the proper disclosures.

Blockstream's CEO Adam Back recently likened many of the biggest altcoin projects to a Ponzi scheme. He does not appear to mind Mow's position at INX, however.

See the original post:
Bitcoin 'maximalists' accused of 'shilling' an SEC-cleared token - Cointelegraph

Read More..

UT Austin Selected as Home of National AI Institute Focused on Machine Learning – UT News | The University of Texas at Austin

AUSTIN, Texas The National Science Foundation has selected The University of Texas at Austin to lead the NSF AI Institute for Foundations of Machine Learning, bolstering the universitys existing strengths in this emerging field. Machine learning is the technology that drives AI systems, enabling them to acquire knowledge and make predictions in complex environments. This technology has the potential to transform everything from transportation to entertainment to health care.

UT Austin already among the worlds top universities for artificial intelligence is poised to develop entirely new classes of algorithms that will lead to more sophisticated and beneficial AI technologies. The university will lead a larger team of researchers that includes the University of Washington, Wichita State University and Microsoft Research.

This is another important step in our universitys ascension as a world leader in machine learning and tech innovation as a whole, and I am grateful to the National Science Foundation for their profound support, said UT Austin interim President Jay Hartzell. Many of the worlds greatest problems and challenges can be solved with the assistance of artificial intelligence, and its only fitting, given UTs history of accomplishment in this area along with the booming tech sector in Austin, that this new NSF institute be housed right here on the Forty Acres.

UT Austin is simultaneously establishing a permanent base for campuswide machine learning research called the Machine Learning Laboratory. It will house the new AI institute and bring together computer and data scientists, mathematicians, roboticists, engineers and ethicists to meet the institutes research goals while also working collaboratively on other interdisciplinary projects. Computer science professor Adam Klivans, who led the effort to win the NSF AI institute competition, will direct both the new institute and the Machine Learning Lab. Alex Dimakis, associate professor of electrical and computer engineering, will serve as the AI institutes co-director.

Machine learning can be used to predict which of thousands of recently formulated drugs might be most effective as a COVID-19 therapeutic, bypassing exhaustive laboratory trial and error, Klivans said. Modern datasets, however, are often diffuse or noisy and tend to confound current techniques. Our AI institute will dig deep into the foundations of machine learning so that new AI systems will be robust to these challenges.

Additionally, many advanced AI applications are limited by computational constraints. For example, algorithms designed to help machines recognize, categorize and label images cant keep up with the massive amount of video data that people upload to the internet every day, and advances in this field could have implications across multiple industries.

Dimakis notes that algorithms will be designed to train video models efficiently. For example, Facebook, one of the AI institutes industry partners, is interested in using these algorithms to make its platform more accessible to people with visual impairments. And in a partnership with Dell Medical School, AI institute researchers will test these algorithms to expedite turnaround time for medical imaging diagnostics, possibly reducing the time it takes for patients to get critical assessments and treatment.

The NSF is investing more than $100 million in five new AI institutes nationwide, including the $20 million project based at UT Austin to advance the foundations of machine learning.

In addition to Facebook, Netflix, YouTube, Dell Technologies and the city of Austin have signed on to transfer this research into practice.

The institute will also pursue the creation of an online masters degree in AI, along with undergraduate research programming and online AI courses for high schoolers and working professionals.

Austin-based tech entrepreneurs Zaib and Amir Husain, both UT Austin alumni, are supporting the new Machine Learning Laboratory with a generous donation to sustain its long-term mission.

The universitys strengths in computer science, engineering, public policy, business and law can help drive applications of AI, Amir Husain said. And Austins booming tech scene is destined to be a major driver for the local and national economy for decades to come.

The Machine Learning Laboratory is based in the Department of Computer Science and is a collaboration among faculty, researchers and students from across the university, including Texas Computing; Texas Robotics; the Department of Statistics and Data Sciences; the Department of Mathematics; the Department of Electrical and Computer Engineering; the Department of Information, Risk & Operations Management; the School of Information; the Good Systems AI ethics grand challenge team; the Oden Institute for Computational Engineering and Sciences; and the Texas Advanced Computing Center (TACC).

View post:
UT Austin Selected as Home of National AI Institute Focused on Machine Learning - UT News | The University of Texas at Austin

Read More..

Participation-washing could be the next dangerous fad in machine learning – MIT Technology Review

More promising is the idea of participation as justice. Here, all members of the design process work together in tightly coupled relationships with frequent communication. Participation as justice is a long-term commitment that focuses on designing products guided by people from diverse backgrounds and communities, including the disability community, which has long played a leading role here. This concept has social and political importance, but capitalist market structures make it almost impossible to implement well.

Machine learning extends the tech industrys broader priorities, which center on scale and extraction. That means participatory machine learning is, for now, an oxymoron. By default, most machine-learning systems have the ability to surveil, oppress, and coerce (including in the workplace). These systems also have ways to manufacture consentfor example, by requiring users to opt in to surveillance systems in order to use certain technologies, or by implementing default settings that discourage them from exercising their right to privacy.

Given that, its no surprise that machine learning fails to account for existing power dynamics and takes an extractive approach to collaboration. If were not careful, participatory machine learning could follow the path of AI ethics and become just another fad thats used to legitimize injustice.

How can we avoid these dangers? There is no simple answer. But here are four suggestions:

Recognize participation as work. Many people already use machine-learning systems as they go about their day. Much of this labor maintains and improves these systems and is therefore valuable to the systems owners. To acknowledge that, all users should be asked for consent and provided with ways to opt out of any system. If they chose to participate, they should be offered compensation. Doing this could mean clarifying when and how data generated by a users behavior will be used for training purposes (for example, via a banner in Google Maps or an opt-in notification). It would also mean providing appropriate support for content moderators, fairly compensating ghost workers, and developing monetary or nonmonetary reward systems to compensate users for their data and labor.

Make participation context specific. Rather than trying to use a one-size-fits-all approach, technologists must be aware of the specific contexts in which they operate. For example, when designing a system to predict youth and gang violence, technologists should continuously reevaluate the ways in which they build on lived experience and domain expertise, and collaborate with the people they design for. This is particularly important as the context of a project changes over time. Documenting even small shifts in process and context can form a knowledge base for long-term, effective participation. For example, should only doctors be consulted in the design of a machine-learning system for clinical care, or should nurses and patients be included too? Making it clear why and how certain communities were involved makes such decisions and relationships transparent, accountable, and actionable.

Plan for long-term participation from the start. People are more likely to stay engaged in processes over time if theyre able to share and gain knowledge, as opposed to having it extracted from them. This can be difficult to achieve in machine learning, particularly for proprietary design cases. Here, its worth acknowledging the tensions that complicate long-term participation in machine learning, and recognizing that cooperation and justice do not scale in frictionless ways. These values require constant maintenance and must be articulated over and over again in new contexts.

Learn from past mistakes. More harm can be done by replicating the ways of thinking that originally produced harmful technology. We as researchers need to enhance our capacity for lateral thinking across applications and professions. To facilitate that, the machine-learning and design community could develop a searchable database to highlight failures of design participation (such as Sidewalk Labs waterfront project in Toronto). These failures could be cross-referenced with socio-structural concepts (such as issues pertaining to racial inequality). This database should cover design projects in all sectors and domains, not just those in machine learning, and explicitly acknowledge absences and outliers. These edge cases are often the ones we can learn the most from.

Its exciting to see the machine-learning community embrace questions of justice and equity. But the answers shouldnt bank on participation alone. The desire for a silver bullet has plagued the tech community for too long. Its time to embrace the complexity that comes with challenging the extractive capitalist logic of machine learning.

Mona Sloane is a sociologist based at New York University. She works on design inequality in the context of AI design and policy.

Excerpt from:
Participation-washing could be the next dangerous fad in machine learning - MIT Technology Review

Read More..

The Role of Artificial Intelligence and Machine Learning in the… – Insurance CIO Outlook

Machine learning has proven to be useful for insurance agents and brokers in various ways. These include capturing knowledge, skills, and expertise from a generation of insurance staff before they retire in the next 5 to 10 years and use it to train new employees.

FREMONT, CA: Technology has become the dominant force across all businesses in the last few years. Disruptive technologies like Artificial Intelligence (AI), machine learning, and natural language processing are improving rapidly and quickly, evolving from theoretical to practical applications. These technologies have also made an impact on insurance agents and brokers. Many people continue to view technology as their foe. They either believe that machines will eventually replace them, or that a machine can never do their job better than them. While this may not be true, some aspects of it are relatable. For instance, a machine will never be able to provide real-time advice as a live agent does. However, low cost and easy to use platforms are currently available that allow agents and brokers to take advantage of this technology to enhance their delivery of advice and expertise to prospects and clients.

Machine learning has proven to be useful for insurance agents and brokers in various ways. These include capturing knowledge, skills, and expertise from a generation of insurance staff before they retire in the next 5 to 10 years and use it to train new employees.

Employee Augmentation

It helps provide personalized answers for a wide range of insurance questions. Digital customers want to get answers for their questions anytime and not just when an agent's office is open.

Personalized Digital Answers

It helps create and deliver a digital annual account review for personal lines or small commercial insurance accountants. A robust analysis leads to client satisfaction, creates cross-selling opportunities, and reduces errors and omission problems for the agency.

Digital Account Review

Many believe that artificial intelligence and machine learning will be the end of insurance agents as a trusted source for adequate protection against financial losses. However, these technologies are a threat only for insurance agents that are simply order takers. Insurance agents and brokers that embrace the technologies will always find opportunities to grow.

These emerging technologies mustn't be seen as a bane but as a boon. Insurance agents and brokers need to work in tandem with the upgrades in technology and leverage it to the best use. It holds increased potential to enhance customer satisfaction and offer a higher quality of service.

Read the original post:
The Role of Artificial Intelligence and Machine Learning in the... - Insurance CIO Outlook

Read More..

Air Force Taps Machine Learning to Speed Up Flight Certifications – Nextgov

Machine learning is transforming the way an Air Force office analyzes and certifies new flight configurations.

The Air Force SEEK EAGLE Office sets standards for safe flight configurations by testing and looking at historical data to see how different storeslike a weapon system attached to an F-16affect flight. A project AFSEO developed along with industry partners can now automate up to 80% of requests for analysis, according to the offices Chief Data Officer Donna Cotton.

The application is kind of like an eager junior engineer consulting a senior engineer, Cotton said. It makes the straightforward calls without any input, but in the hard cases it walks into the senior engineers office and says: Hey, I did a bunch of research and this is what I found out. Can you give me your opinion?

Cotton spoke at a Tuesday webinar hosted by Tamr, one of the industry partners involved in the project. Tamr announced July 30 AFSEO awarded the company a $60 million contract for its machine learning application. Two other companies, Dell and Cloudera, helped AFSEO take decades of historical data from simulations, performance studies and the like that were siloed across various specialities and organize them into a searchable data lake.

On top of this new data architecture, the machine learning application provided by Tamr searches through all the historical data to find past records that can help answer new safety recommendation requests automatically.

This tool is critical because the vast majority of AFSEOs flight certification recommendations are made by analogy, meaning using previous data rather than new flight tests. But in the past, data was disorganized and lacked unification. This made tracking down these helpful records a challenge for engineers.

Now, a cleaner AFSEO data lake cuts the amount of time engineers waste on looking for the information they need. Machine learning further speeds up the process by generating safety reports automatically while still keeping the professional engineers in the loop. Even when engineers need to produce original research, the machine learning application can smooth the process by collecting related records to serve as a jumping off point.

The new process helps AFSEO avoid doing costly flight tests while also increasing confidence that the team is making the safety certification correctly with all the information available to them, Cotton said.

We are able to be more productive, Cotton said. It's saving us a lot of money because for us, it's not about profit, but it's about hours. It's about how much effort are we going to have to use to solve or to answer a new request.

See original here:
Air Force Taps Machine Learning to Speed Up Flight Certifications - Nextgov

Read More..

Machine Learning Artificial intelligence Market Size and Growth By Leading Vendors, By Types and Application, By End Users and Forecast to 2020-2027 -…

New Jersey, United States,- Market Research Intellect recently published a report on the Machine Learning Artificial intelligence Market. The study was supported by data obtained either from primary sources or from corporate databases. The experts in the market have confirmed that the data is realistic and relevant to the particular market conditions and therefore will prove extremely helpful to the user. The factors that have been broken down into driver and restraint systems. The regions, types, applications, and strategies are segmented and subdivided for better and better understanding.

This report covers the current economic impact of COVID-19. This outbreak drastically changed the global economic situation. The current scenario of the constantly evolving corporate sector, as well as the present and future assessment of the impact, are also addressed in the report.

The Machine Learning Artificial intelligence marketreport gives a 360 approach for a holistic understanding of the market scenario. It relies on authentically-sourced information and an industry-wide analysis to predict the future growth of the sector. The study gives a comprehensive assessment of the Machine Learning Artificial intelligence industry, along with market segmentation, product types, applications, and value chain.

Leading Machine Learning Artificial intelligence manufacturers/companies operating at both regional and global levels:

The report also inspects the financial standing of the leading companies, which includes gross profit, revenue generation, sales volume, sales revenue, manufacturing cost, individual growth rate, and other financial ratios.

Research Objective:

Our panel of trade analysts has taken immense efforts in doing this group action in order to produce relevant and reliable primary & secondary data regarding the Machine Learning Artificial intelligence market. Also, the report delivers inputs from the trade consultants that will help the key players in saving their time from the internal analysis. Readers of this report are going to be profited with the inferences delivered in the report. The report gives an in-depth and extensive analysis of the Machine Learning Artificial intelligence market.

The Machine Learning Artificial intelligence Market is Segmented:

In market segmentation by types of Machine Learning Artificial intelligence, the report covers-

In market segmentation by applications of the Machine Learning Artificial intelligence, the report covers the following uses-

This Machine Learning Artificial intelligence report umbrellas vital elements such as market trends, share, size, and aspects that facilitate the growth of the companies operating in the market to help readers implement profitable strategies to boost the growth of their business. This report also analyses the expansion, market size, key segments, market share, application, key drivers, and restraints.

Machine Learning Artificial intelligence Market Regional Analysis:

Geographically, the Machine Learning Artificial intelligence market is segmented across the following regions:North America, Europe, Latin America, Asia Pacific, and Middle East & Africa.

Key Coverage of Report:

Key insights of the report:

In conclusion, the Machine Learning Artificial intelligence Market report provides a detailed study of the market by taking into account leading companies, present market status, and historical data to for accurate market estimations, which will serve as an industry-wide database for both the established players and the new entrants in the market.

About Us:

Market Research Intellect provides syndicated and customized research reports to clients from various industries and organizations with the aim of delivering functional expertise. We provide reports for all industries including Energy, Technology, Manufacturing and Construction, Chemicals and Materials, Food and Beverage, and more. These reports deliver an in-depth study of the market with industry analysis, the market value for regions and countries, and trends that are pertinent to the industry.

Contact Us:

Mr. Steven Fernandes

Market Research Intellect

New Jersey ( USA )

Tel: +1-650-781-4080

See the original post here:
Machine Learning Artificial intelligence Market Size and Growth By Leading Vendors, By Types and Application, By End Users and Forecast to 2020-2027 -...

Read More..

Chatbots Are Machine Learning Their Way To Human Language – Forbes

Moveworks founding team from left to right Vaibhav Nivargi, CTO; Bhavin Shah, CEO; Varun Singh, VP ... [+] of Product; Jiang Chen, VP of Machine Learning.

Computers and humans have never spoken the same language. Over and above speech recognition, we also need computers to understand the semantics of written human language. We need this capability because we are building the Artificial Intelligence (AI)-powered chatbots that now form the intelligence layers in Robot Process Automation (RPA) systems and beyond.

Known formally as Natural Language Understanding (NLU), early attempts (as recently as the 1980s) to give computers the ability to interpret human text were comically terrible. This was a huge frustration to both the developers attempting to make these systems work and the users exposed to these systems.

Computers are brilliant at long division, but really bad at knowing the difference between whether humans are referring to football divisions, parliamentary division lobbies or indeed long division for mathematics. This is because mathematics is formulaic, universal and unchanging, but human language is ambiguous, contextual and dynamic.

As a result, comprehending a typical sentence requires the unprogrammable quality of common sense or so we thought.

But in just the last few years, software developers in the field of Natural Language Understanding (NLU) have made several decades worth of progress in overcoming that obstacle, reducing the language barrier between people and AI by solving semantics with mathematics.

Such progress has stemmed in no small part from giant leaps forward in NLU models, including the landmark BERT framework and offshoots like DistilBERT, RoBERTa and ALBERT. Powered by hundreds of these models, modern NLU software is able to deconstruct complex sentences to distill their essential meaning, said Vaibhav Nivargi, CTO and co-founder of Moveworks.

Moveworks software combines AI with Natural Language Processing (NLP) to understand and interpret user requests, challenges and problems before then using a further degree of AI to help deliver the appropriate actions to satisfy the users needs.

Nivargi explains that crucially here we can also now build chatbots that use Machine Learning (ML) to go a step further: autonomously addressing users requests and troubleshooting questions written in natural language. So not only can AI now communicate with employees on their terms, it can even automate many of the routine tasks that make work feel like work - thanks to this newfound capacity for reading comprehension.

Nivargi provides an illustrative example of an IT support request, which we can break down and analyze. Bhavin is a new company employee and a user is asking the chatbot how he can be added to the organizations marketing group to access its information pool and data. The request is as follows (graphic shown below at end):

Howdo [sic] I add Bhavin to the marketing group.

In large part due to the typing/spelling mistake at the start (instead of how do, the user has typed howdo) we have an immediate problem. As recently as two years ago, there was not a single application in the world capable of understanding (and then resolving) the infinite variety of similar requests to this that employees pose to their IT teams.

Of course, we could program an application to trigger the right automated workflow when it receives this exact request. But needless to say, that approach doesnt scale at all. Hard problems demand hard solutions. So here, any solution worth its salt must tackle the fundamental challenges of natural language, which is ambiguous, contextual and dynamic, said Nivargi.

A single word can have many possible meanings; for instance, the word run has about 645 different definitions. Add in the inevitable human error like the typo in this request of the phrase how do and we can see that breaking down a single sentence becomes quite daunting, quite quickly. MoveworksNivargi explains that the initial step, therefore, is to use machine learning to identify syntactic structures that can help us rectify spelling or grammatical errors.

But, he says, to disambiguate what the employee wants, we also need to consider the context surrounding their request, including that employees department, location and role, as well as other relevant entities. A key technique in doing so is meta learning, which entails analyzing so-called metadata (information about information).

By probabilistically weighing the fact that Alex (another employee) and Bhavin are located in North America, Machine Learning models can fuzzy select the marketingna@company.abc email group, without Alex having to have specified his or her exact name. In this way, we can potentially get Alexs help and get him/her involved in the workflow at hand, said Nivargi.

As TechTarget explains here, Fuzzy logic is an approach to computing based on degrees of truth rather than the usual "true or false" (1 or 0) Boolean logic on which the modern computer is based.

Human service desk agents already factor in context by drawing on their experience, so the secret for an AI chatbot is to mimic this intuition with mathematical models.

Finally lets remember that language in particular the language used in the enterprise is dynamic. New words and expressions arise every month, while the IT systems and applications at a given company shift even more often. To deal with so much change, an effective chatbot must be rooted in advanced Machine Learning, since it needs to constantly retrain itself based on real-time information.

Despite the complexity under the hood, however, the number one criteria for a successful chatbot is a seamless user experience. Nivargi says that what his firm has learned when developing NLU technologies is that all employees care about is getting their requests resolved, instantly, via natural conversations on a messaging tool.

As we stand at the turn of the decade, we humans are arguably still not 100% comfortable with chatbot interactions. Theyre still too automated, too often non-intuitive and (perhaps unsurprisingly) too to machine-like. Technologies like these show that we've started to build chatbots with semantic intuitive intelligence, but there is still work to do. When we get to a point where technology can navigate the peculiarities and idiosyncrasies of human language.... then, just then, we may start to enjoy talking to robots.

Addressing requests written in natural language requires the combination of hundreds of machine ... [+] learning models. In this case, the Moveworks chatbot determines that Alex wants to add Bhavin to the email group for marketing.

Read the original:
Chatbots Are Machine Learning Their Way To Human Language - Forbes

Read More..

What is AutoML and Why Should Your Business Consider It – BizTech Magazine

Automation offers substantive benefits as companies look for ways to manage evolving workforces and workplace expectations. More than half of U.S. businesses now plan to increase their automation investment to help increase their agility and improve their ability to handle changing conditions quickly, according to Robotics and Automation News.

Businesses also need to be able to solve problems at scale, something that organizations are increasingly turning to machine learning to do. By creating algorithms that learn over time, its possible for companies to streamline decision-making with data-driven predictions. But creating the models can be complex and time-consuming, putting an added strain on businesses that may be low on resources.

Automated machine learning combines these two technologies to tap the best of both worlds, allowing companies to gain actionable insights while reducing total complexity. Once implemented, AutoML can help businesses gather and analyze data, respond to it quickly and better manage resources.

WATCH: Find out how organizations can empower digital transformation and secure remote work.

AutoML goes a step further than classic machine learning, says Earnest Collins, managing member of Regulatory Compliance and Examination Consultants and a member of the ISACA Emerging Technologies Advisory Group.

AutoML goes beyond creating machine learning architecture models, says Collins. It can automate many aspects of machine learning workflow, which include data preprocessing, feature engineering, model selection, architecture search and model deployment.

AutoML deployments can also be categorized by the format of training data used. Collins points to examples such as independent, identically distributed (IID) tabular data, raw text or image data, and notes that some AutoML solutions can handle multiple data types and algorithms.

There is no single algorithm that performs best on all data sets, he says.

Leveraging AutoML solutions offers multiple benefits that go beyond traditional machine learning or automation. The first is speed, according to Collins.

AutoML allows data scientists to build a machine learning model with a high degree of automation more quickly and conduct hyperparameter search over different types of algorithms, which can otherwise be time-consuming and repetitive, he says. By automating key processes from raw data set capture to eventual analysis and learningteams can reduce the amount of time required to create functional models.

Another benefit is scalability. While machine learning models cant compete with the in-depth nature of human cognition, evolving technology makes it possible to create effective analogs of specific human learning processes. Introducing automation, meanwhile, helps apply this process at scale in turn, enabling data scientists, engineers and DevOps teams to focus on business problems instead of iterative tasks, Collins says.

A third major benefit is simplicity, according to Collins. AutoML is a tool that assists in automating the process of applying machine learning to real-world problems, he says.

By reducing the complexity that comes with building, testing and deploying entirely new ML frameworks, AutoML streamlines the processes required to solve line-of-business challenges.

For machine learning solutions to deliver business value, ML models must be optimized based on current conditions and desired outputs. Doing so requires the use of hyperparameters, which Collins defines as adjustable parameters that govern the training of ML models.

Optimal ML model performance depends on the hyperparameter configuration value selection; this can be a time-consuming, manual process, which is where AutoML can come into play, Collins adds.

By using AutoML platforms to automate key hyperparameter selection and balancing including learning rate, batch size and drop rate its possible to reduce the amount of time and effort required to get ML algorithms up and running.

While AutoML isnt new, evolution across machine learning and artificial intelligence markets is now driving a second generation of automated machine learning platforms, according to RTInsights. The first wave of AutoML focused on building and validating models, but the second iterations include key features such as data preparation and feature engineering to accelerate data science efforts.

But this market remains both fragmented and complex, according to Forbes, because of a lack of established standards and expectations in the data science and machine learning (DSML) industry. Businesses can go with an established provider, such as Microsoft Azure Databricks, or they can opt for more up-and-coming solutions such as Google Cloud AutoML.

There are more tools around the corner. According to Synced, Google researchers are now developing AutoML-Zero, which is capable of searching for applicable ML algorithms within a defined space to reduce the need to create them from scratch. The search giant is also applying its AutoML to unique use cases; for example, the companys new Fabricius tool which leverages Googles AutoML vision toolset is designed to decode ancient Egyptian hieroglyphics.

Technological advancements combined with shifting staff priorities are somewhat driving robotic replacements. According to Time, companies are replacing humans wherever possible to reduce risk and improve operational output. But that wont necessarily apply to data scientists as AutoML rises, according to Collins.

The skills of professional, well-trained data scientists will be essential to interpreting data and making recommendations for how information should be used, he says. AutoML will be a key tool for improving their productivity, and the citizen data scientist, with no training in the field, would not be able to do machine learning without AutoML.

In other words, while AutoML platforms provide business benefits, recognizing the full extent of automated advantages will always require human expertise.

See the original post here:
What is AutoML and Why Should Your Business Consider It - BizTech Magazine

Read More..

Explainable AI: From the peak of inflated expectations to the pitfalls of interpreting machine learning models – ZDNet

Machine learning and artificial intelligence are helping automate an ever-increasing array of tasks, with ever-increasing accuracy. They are supported by the growing volume of data used to feed them, and the growing sophistication in algorithms.

The flip side of more complex algorithms, however, is less interpretability. In many cases, the ability to retrace and explain outcomes reached by machine learning models (ML) is crucial, as:

"Trust models based on responsible authorities are being replaced by algorithmic trust models to ensure privacy and security of data, source of assets and identity of individuals and things. Algorithmic trust helps to ensure that organizations will not be exposed to the risk and costs of losing the trust of their customers, employees and partners. Emerging technologies tied to algorithmic trust include secure access service edge, differential privacy, authenticated provenance, bring your own identity, responsible AI and explainable AI."

The above quote is taken from Gartner's newly released 2020 Hype Cycle for Emerging Technologies. In it, explainable AI is placed at the peak of inflated expectations. In other words, we have reached peak hype for explainable AI. To put that into perspective, a recap may be useful.

As experts such as Gary Marcus point out, AI is probably not what you think it is. Many people today conflate AI with machine learning. While machine learning has made strides in recent years, it's not the only type of AI we have. Rule-based, symbolic AI has been around for years, and it has always been explainable.

Incidentally, that kind of AI, in the form of "Ontologies and Graphs" is also included in the same Gartner Hype Cycle, albeit in a different phase -- the trough of disillusionment. Incidentally, again, that's conflating.Ontologies are part of AI, while graphs, not necessarily.

That said: If you are interested in getting a better understanding of the state of the art in explainable AI machine learning, reading Christoph Molnar's book is a good place to start. Molnar is a data scientist and Ph.D. candidate in interpretable machine learning. Molnar has written the bookInterpretable Machine Learning: A Guide for Making Black Box Models Explainable, in which he elaborates on the issue and examines methods for achieving explainability.

Gartner's Hype Cycle for Emerging Technologies, 2020. Explainable AI, meaning interpretable machine learning, is at the peak of inflated expectations. Ontologies, a part of symbolic AI which is explainable, is in the trough of disillusionment

Recently, Molnar and a group of researchers attempted to addresses ML practitioners by raising awareness of pitfalls and pointing out solutions for correct model interpretation, as well as ML researchers by discussing open issues for further research. Their work was published as a research paper, titledPitfalls to Avoid when Interpreting Machine Learning Models, by the ICML 2020 Workshop XXAI: Extending Explainable AI Beyond Deep Models and Classifiers.

Similar to Molnar's book, the paper is thorough. Admittedly, however, it's also more involved. Yet, Molnar has striven to make it more approachable by means of visualization, using what he dubs "poorly drawn comics" to highlight each pitfall. As with Molnar's book on interpretable machine learning, we summarize findings here, while encouraging readers to dive in for themselves.

The paper mainly focuses on the pitfalls of global interpretation techniques when the full functional relationship underlying the data is to be analyzed. Discussion of "local" interpretation methods, where individual predictions are to be explained, is out of scope. For a reference on global vs. local interpretations, you can refer to Molnar's book as previously covered on ZDNet.

Authors note that ML models usually contain non-linear effects and higher-order interactions. As interpretations are based on simplifying assumptions, the associated conclusions are only valid if we have checked that the assumptions underlying our simplifications are not substantially violated.

In classical statistics this process is called "model diagnostics," and the research claims that a similar process is necessary for interpretable ML (IML) based techniques. The research identifies and describes pitfalls to avoid when interpreting ML models, reviews (partial) solutions for practitioners, and discusses open issues that require further research.

Under- or overfitting models will result in misleading interpretations regarding true feature effects and importance scores, as the model does not match the underlying data generating process well. Evaluation of training data should not be used for ML models due to the danger of overfitting. We have to resort to out-of-sample validation such as cross-validation procedures.

Formally, IML methods are designed to interpret the model instead of drawing inferences about the data generating process. In practice, however, the latter is the goal of the analysis, not the former. If a model approximates the data generating process well enough, its interpretation should reveal insights into the underlying process. Interpretations can only be as good as their underlying models. It is crucial to properly evaluate models using training and test splits -- ideally using a resampling scheme.

Flexible models should be part of the model selection process so that the true data-generating function is more likely to be discovered. This is important, as the Bayes error for most practical situations is unknown, and we cannot make absolute statements about whether a model already fits the data optimally.

Using opaque, complex ML models when an interpretable model would have been sufficient (i.e., having similar performance) is considered a common mistake. Starting with simple, interpretable models and gradually increasing complexity in a controlled, step-wise manner, where predictive performance is carefully measured and compared is recommended.

Measures of model complexity allow us to quantify the trade-off between complexity and performance and to automatically optimize for multiple objectives beyond performance. Some steps toward quantifying model complexity have been made. However, further research is required as there is no single perfect definition of interpretability but rather multiple, depending on the context.

This pitfall is further analyzed in three sub-categories: Interpretation with extrapolation, confusing correlation with dependence, and misunderstanding conditional interpretation.

Interpretation with Extrapolation refers to producing artificial data points that are used for model predictions with perturbations. These are aggregated to produce global interpretations. But if features are dependent, perturbation approaches produce unrealistic data points. In addition, even if features are independent, using an equidistant grid can produce unrealistic values for the feature of interest. Both issues can result in misleading interpretations.

Before applying interpretation methods, practitioners should check for dependencies between features in the data (e.g., via descriptive statistics or measures of dependence). When it is unavoidable to include dependent features in the model, which is usually the case in ML scenarios, additional information regarding the strength and shape of the dependence structure should be provided.

Confusing correlation with dependence is a typical error. The Pearson correlation coefficient (PCC) is a measure used to track dependency among ML features. But features with PCC close to zero can still be dependent and cause misleading model interpretations. While independence between two features implies that the PCC is zero, the converse is generally false.

Any type of dependence between features can have a strong impact on the interpretation of the results of IML methods. Thus, knowledge about (possibly non-linear) dependencies between features is crucial. Low-dimensional data can be visualized to detect dependence. For high-dimensional data, several other measures of dependence in addition to PCC can be used.

Misunderstanding conditional interpretation. Conditional variants to estimate feature effects and importance scores require a different interpretation. While conditional variants for feature effects avoid model extrapolations, these methods answer a different question. Interpretation methods that perturb features independently of others also yield an unconditional interpretation.

Conditional variants do not replace values independently of other features, but in such a way that they conform to the conditional distribution. This changes the interpretation as the effects of all dependent features become entangled. The safest option would be to remove dependent features, but this is usually infeasible in practice.

When features are highly dependent and conditional effects and importance scores are used, the practitioner has to be aware of the distinct interpretation. Currently, no approach allows us to simultaneously avoid model extrapolations and to allow a conditional interpretation of effects and importance scores for dependent features.

Global interpretation methods can produce misleading interpretations when features interact. Many interpretation methods cannot separate interactions from main effects. Most methods that identify and visualize interactions are not able to identify higher-order interactions and interactions of dependent features.

There are some methods to deal with this, but further research is still warranted. Furthermore, solutions lack in automatic detection and ranking of all interactions of a model as well as specifying the type of modeled interaction.

Due to the variance in the estimation process, interpretations of ML models can become misleading. When sampling techniques are used to approximate expected values, estimates vary, depending on the data used for the estimation. Furthermore, the obtained ML model is also a random variable, as it is generated on randomly sampled data and the inducing algorithm might contain stochastic components as well.

Hence, themodel variance has to be taken into account. The true effect of a feature may be flat, but purely by chance, especially on smaller data, an effect might algorithmically be detected. This effect could cancel out once averaged over multiple model fits. The researchers note the uncertainty in feature effect methods has not been studied in detail.

It's a steep fall to the peak of inflated expectations to the trough of disillusionment. Getting things done for interpretable machine learning takes expertise and concerted effort.

Simultaneously testing the importance of multiple features will result in false-positive interpretations if the multiple comparisons problem (MCP) is ignored. MCP is well known in significance tests for linear models and similarly exists in testing for feature importance in ML.

For example, when simultaneously testing the importance of 50 features, even if all features are unimportant, the probability of observing that at least one feature is significantly important is 0.923. Multiple comparisons will even be more problematic, the higher dimensional a dataset is. Since MCP is well known in statistics, the authors refer practitioners to existing overviews and discussions of alternative adjustment methods.

Practitioners are often interested in causal insights into the underlying data-generating mechanisms, which IML methods, in general, do not provide. Common causal questions include the identification of causes and effects, predicting the effects of interventions, and answering counterfactual questions. In the search for answers, researchers can be tempted to interpret the result of IML methods from a causal perspective.

However, a causal interpretation of predictive models is often not possible. Standard supervised ML models are not designed to model causal relationships but to merely exploit associations. A model may, therefore, rely on the causes and effects of the target variable as well as on variables that help to reconstruct unobserved influences.

Consequently, the question of whether a variable is relevant to a predictive model does not directly indicate whether a variable is a cause, an effect, or does not stand in any causal relation to the target variable.

As the researchers note, the challenge of causal discovery and inference remains an open key issue in the field of machine learning. Careful research is required to make explicit under which assumptions what insight about the underlying data generating mechanism can be gained by interpreting a machine learning model

Molnar et. al. offer an involved review of the pitfalls of global model-agnostic interpretation techniques for ML. Although as they note their list is far from complete, they cover common ones that pose a particularly high risk.

They aim to encourage a more cautious approach when interpreting ML models in practice, to point practitioners to already (partially) available solutions, and to stimulate further research.

Contrasting this highly involved and detailed groundwork to high-level hype and trends on explainable AI may be instructive.

Read more:
Explainable AI: From the peak of inflated expectations to the pitfalls of interpreting machine learning models - ZDNet

Read More..

Focusing on ethical AI in business and government – FierceElectronics

The World Economic Forum and associate partner Appen are wrestling with the thorny issue of how to create artificial intelligence with a sense of ethics.

Their main area of focus is to design standards and best practices for responsible training data used in building machine learning and AI applications. It has already been a long process and continues.

A solid training data platform and management strategy is often the most critical component of launching a successful, responsible machine learning-powered product into production, said Mark Brayan, CEO of Appen in a statement. Appen has been providing training data to companies building AI for more than 20 years. In 2019, Appen created its own Crowd Code of Ethics.

Interesting read? Subscribe to FierceElectronics!

The electronics industry remains in flux as constant innovation fuels market trends. FierceElectronics subscribers rely on our suite of newsletters as their must-read source for the latest news, developments and predictions impacting their world. Sign up today to get electronics news and updates delivered to your inbox and read on the go.

Ethical, diverse training data is essential to building a responsible AI system, Brayan added.

Kay Firth-Butterfield, head of AI and machine learning at WEF, said the industry needs guidelines for acquiring and using responsible training data. Companies should address questions around user permissions, privacy, security, bias, safety and how people are compensated for their work in the AI supply chain, she said.

Every business needs a plan to understand AI and deploy AI safely and ethically, she added in a video overview of Forums AI agenda. The purpose is to think about what are the big issues in AI that really require something be done in the governance area so that AI can flourish.

Were very much advocating asoft law approach, thinking about standards and guidelines rather than looking to regulation, she said.

The Forum has issued a number of white papers dating to 2018 on ethics and related topics, with a white paper on responsible limits on facial recognition issued in March.

RELATED: Researchers deploy AI to detect bias in AI and humans

In January, the Forum published its AI toolkit for boards of directors with 12 modules for the impacts and potential of AI in company strategy and is currently building a toolkit for transferring those insights to CEOs and other C-suite executives.

Another focus area is on human-centered AI for human resources to create a toolkit for HR professionals that will help promote ethical human-centered use of AI. Various HR tools have been developed in recent years that rely on AI to hire and retain talent and the Forum notes that concerns have been raised about AI algorithms encoding bias and discrimination. Errors in the adoption of AI-based products can also undermine employee trust, leading to lower productivity and job satisfaction, the Forum added.

Firth-Butterfield will be a keynote speaker at Appen annual Train AI conference on October 14.

RELATED: Tech firms grapple with diversity after George Floyd protests

Read the original:
Focusing on ethical AI in business and government - FierceElectronics

Read More..