Page 1,123«..1020..1,1221,1231,1241,125..1,1301,140..»

DeepMind Alum Wants to Use AI to Speed the Development of … – Data Center Knowledge

(Bloomberg) -- Ever since ChatGPT went virallast fall, companieshave touted many waysartificial intelligencecan make ourlives easier. Theyve promised superhuman virtual assistants, tutors, lawyers and doctors.

What about a superhuman chemical engineer?

Related: Data Center Sustainability: Green Solutions for the Future

London-based startupOrbital Materials would like tocreate just that. The startup is working to applygenerative AI the method behind tools like ChatGPT expresslyforaccelerating the development ofclean energytechnologies. Essentially, the idea is to make computer models powerful and sharp enough to identify the best formulas for products likesustainable jet fuelor batteries free of rare-earth minerals.

Jonathan Godwin, anOrbital Materials co-founder, imagines a system thats as accessible and effective as the software engineers use today to model designs for things likeairplane wings and household furniture.

Related: HPE Unveils AI Supercomputer Cloud Service

That, historically, has just been too difficult for molecular science, he said.

ChatGPT works because its adept at predicting text heres the next word or sentence that makes sense. For the same idea to work in chemistry, an AI system would need to predict how a new molecule would behave, not just in a lab but in the real world.

Several researchers and companies have deployedAI to hunt for newer, greener materials. Symyx Technologies, a materials discovery company formed in 1990s,wound down after a sale. More recentcompanies have gained traction makingpetrochemical alternativesandprogramming cells.

Still, for many materials needed to decarbonize the planet, the technology isntthere yet.

It cantake decadesfor a new advanced material to move from discovery to the market. That timeline is way too slow for the businesses and nations looking to rapidly cut emissions as they race to meetnet zero targets.

That needs to happen in the next 10 years, or sooner, said Aaike van Vugt, co-founder of material science startup VSParticle.

AI researchers think they can help.Before launching Orbital Materials, Godwin spent three years researching advanced material discovery atDeepMind, Googles AI lab. That lab releasedAlphaFold, amodel to predictprotein structures that could speed up the search for new drugs and vaccines.That, coupled with the rapidtakeoff of tools like ChatGPT, convinced him that AI would soon be capable of conquering the material world.

What I thought would take 10 years was happening in a matter of 18 months, he said. Things are getting better and better and better.

Godwin compareshis method withOrbital Materials toAI image generators like Dall-E and Stable Diffusion. Those models are created using billions of online images so that when users type in a text prompt, a photorealistic creation appears. Orbital Materials plans to trainmodels with loads of data on the molecular structure ofmaterials. Type in some desired property and material say, an alloy that can withstand very high heat and the model spits out a proposed molecular formula.

In theory, this approach is effective because it can both imagine new moleculesandmeasure howthey will work, said Rafael Gomez-Bombarelli, an assistant professor at MIT, who advisedOrbital Materials. (He said he is not an investor.)

Right now, many tech investors are prowling for companies that can turn a profit byimproving greener material production.Thats particularly the case in Europe, where regulatorsare forcing manufacturersto lower carbon emissions or face stiff fines. The markets for advanced materials in sectors like renewable energy, transportation and agriculture are set togrow by tens of billions of dollars in the coming years.

Some researchers, like those at theUniversity of Toronto, have set up self-driving labs that pair AI systems with robots to search for new materials at unparalleled speeds. Dutch startup VSParticle makes machinery used to develop components for gas sensors and green hydrogen.

Think of it like aDNA sequencer in a genomics lab, said co-founder van Vugt,who believes his equipment can help shorten the 20-year time horizon of advanced materials to one year, and, eventually,a couple of months. His company is currently raising investment capital.

Orbital Materials, which raised $4.8 million in previously undisclosed initial funding, is planning to start withturning its AI gaze towardcarbon capture. The startup is working on an algorithmic model that designsmolecular sieves, ortiny pellets installedwithin a device that can sift CO2 and other noxious chemicals from other emissions,more efficiently than current methods.(Godwin said the startup, which has several AI researchers, plans to publish peer-reviewed results on this tech soon.) Carbon capture has failed to work at scale to date, though thanks to a slew of government incentives,particularly in the US, interest in deploying the technology is rapidly ramping up.

Eventually, Godwin said Orbital Materials would like tomove into areas like fuel and batteries. He imagines mirroring thebusiness model ofsynthetic biology and drug discovery companies: develop the brainpower, then license out the software or novel materials to manufacturers. Its going to take us a little bit of time to get to market," said Godwin. "But once youre there, it happens very quickly.

But getting the AI right is only half the battle. Actually making advanced materialsin areas like battery and fuel production requires working with huge incumbent enterprises and messy supply chains. This can be even costlier than developing new drugs,argued MITs Gomez-Bombarelli.

The economics and de-risking make it just way harder, he said.

Heather Redman, a managing partner with Flying Fish Partners, which backedOrbital Materials, said most tech investors chasing the shiny penny of generative AI have failed to look at its applications outside of chatbots.She acknowledged the risks of startups working in the energy sector, butbelievesthe $1 trillion potential of markets like batteries and carbon capture are worth the investing risk.

We love big hills as long as theres a big gigantic market and opportunity at the top, she said.

Gomez-Bombarelli is aware how big these hills can be. He helped start a similar company to Orbital Materials in 2015, calledCalculario, which used AI and quantum chemistry to speed up the discovery process for a range ofnew materials.It didnt get enough traction and had to focus on the OLED industry.

Maybe we didnt make our case, he said. Or maybe the market wasnt ready.

Whether it is now is an open question. But there are encouraging signs.Computing certainly has improved. Newcomers might also have an easier time selling AI because would-be customers could more easily graspthe potential. Gomez-Bombarelli said the pitch is relatively simple:Look at ChatGPT. Wecan do the same thing for chemistry.

Excerpt from:
DeepMind Alum Wants to Use AI to Speed the Development of ... - Data Center Knowledge

Read More..

5 things about AI you may have missed today: Google to use public info to train AI, tech layoffs, more – HT Tech

After announcing Gemini, a project by DeepMind aiming to surpass artificial intelligence models like ChatGPT yesterday, Google has now updated its privacy policy and is suggesting that it will only be using publicly available data to train its AI models. In other news, there is a growing number of tech layoffs due to the rise of AI which has concerned many researchers who believed tech roles would largely remain safe initially. This and more in todays AI roundup. Let us take a closer look.

A report by Gizmodo has revealed that Google has updated its privacy policy and now suggests that it will use any data that is publicly available (can be read by Google) to train its AI models.

Google uses the information to improve our services and to develop new products, features, and technologies that benefit our users and the public. For example, we use publicly available information to help train Googles AI models and build products and features like Google Translate, Bard, and Cloud AI capabilities, mentions the new policy.

The most important part of this newly updated privacy policy is that earlier Google said that the data will be used for its language models, which have now been replaced with AI models. This raises serious concerns about the privacy of individuals who post things online, as now it is not just about who has access to the data but the users do not even control how the data can be used.

It was speculated that this was one of the reasons why both Reddit and Twitter made drastic policy changes to keep AI data harvesting at bay.

A new report by CNN claims that there is a growing number of layoffs occurring in the tech sector, a majority of which are linked to AI. Many employees have been fired and hiring has been frozen as companies figure out which roles can be taken over by AI.

Highlighting such an instance, the report mentioned IBM CEO Arvind Krishnas statements from an interview with Bloomberg where he mentioned that the company was going to stop its hiring to understand where AIs role can be more impactful.

Meesho and the Vision and AI Lab (VAL) of the Indian Institute of Science (IISc) will be collaborating together to conduct research into generative AI, a report by Business Standard said. The two have also signed a memorandum of understanding (MoU) of one year.

As per the MoU, Meesho will let its data scientists work with researchers from IISc to focus on multimodal representation learning and generative AI capabilities. Meesho believes this collaboration will result in the expansion of the e-retail sector by harnessing the abilities of AI.

The universities in the UK have prepared guiding principles around generative AI. The central focus is to provide education and awareness on such technologies as institutes struggle to adapt teaching and assessment methods to adjust to the rise of AI, a report by The Guardian stated.

Unlike the previous sentiments that AI should be banned from educational institutions, the new guidelines emphasize the need to learn and adapt to this growing technology to tap into its potential while also making the students informed about the risks of plagiarism, biases, and inaccuracies of AI.

An Instagram page by the username ai.magine_ has shared a series of photos showcasing characters from the popular sitcom Friends being reimagined in Indian ethnic dresses. The post showcases Chandler and Monica tying the knot at an Indian wedding. It also shows Joey attending the wedding wearing an Indian sherwani.

More here:
5 things about AI you may have missed today: Google to use public info to train AI, tech layoffs, more - HT Tech

Read More..

Vines in Mind – richmondmagazine.com – Richmond magazine

When David Hunsaker and his wife, Barbara, get on the subject of tomatoes, youd better get comfortable, especially if theyve enjoyed a glass or two of sparkling ros. Known locally as the tomato king and queen, the vivacious duo and enthusiastic leaders of Village Garden farm over 300 varieties of the fruit on their small farm in Hanover that boasts a tomato-friendly terroir a trifecta of clay, dirt and sand.

For the past two years, the couple, along with enthusiastic oenophile and sommelier Jason Tesauro of Barboursville Vineyards, have been calling upon top chefs and food and beverage professionals across the state to elevate the humble tomato through the Summer Supper Somm dinner series. And in 2023, the ambitious ambassadors of the commonwealths bounty are back at it for a third go-round.

Its a good time and great food, its a different interaction, a little bit of a different experience, and thats exactly what were striving for, David says with a wide smile.

The tomato showcase held its first service of the year on June 26 at Shagbark and concludes on Aug. 13 at Zoes Steak & Seafood in Virginia Beach. Other events in the series include an Indian brunch at Lehja in Short Pump, a dinner at pioneering Parisian instiution LAuberge Chez Francois in Northern Virginia, a Church Hill tomato crawl, a walk-around tasting soiree at Lewis Ginter Botanical Garden and backyard party on the Hunsaker farm, in addition to many other juicy destinations in between. Each dinner includes pairings with wines from Barboursville.

The series has been garnering a fan base of returning guests much like concertgoers who get hooked and hop on tour to see their favorite musicians and participating restaurants. Tesauro notes that about one out of every four diners is a repeat attendee, food fanatics who get giddy over the varied and versatile produce just as much as the chefs who are uncovering its potential.

And while the founders of Summer Supper Somm have a fervent dedication to and reverence for the fruit behind their series, that deep-seeded admiration is simply part of the events enticing nature, and its exploration of tomatoes rustic roots and endless possibilities.

When asked how theyll keep things fresh and interesting during this years events, Barbara, who has a soft spot for a ribbed Tlacolula pink, replies without pause, I think the tomato does that for us.

While adhering to the series laid-back flair and not too many guard rails mantra, David says that with this third iteration, theyve learned that communal dining works best, and staggered seating not so much. This year, participating chefs are also encouraged to dig a little deeper, cracking open cookbooks of the past to gain inspiration for dishes that are historically inspired.

The first ketchup, the first tomato gravy, the first tomato aspic, all of these things have history, he says.

Newcomers on the bill for the nearly summerlong ode to tomatoes include 21 Spoons in Midlothian, Magnolias at The Mill in Purcellville, Michelin-starred Marcels by Robert Wiedmaier in Washington D.C., Lewis Ginter Botanical Garden, Acacia Midtown and Yellow Umbrella Provisions. During the Church Hill Tomato Crawl, Sub Rosa Bakery, 8 1/2 andCobra Burger will feature tomato-centric specials on their menu.

During this series, there will be an opportunity for people to have the light come on about whats available in their own neighborhood and out their back door, Tesauro says. These are tomatoes from where we live, this is wine thats grown a corks throw from where youre building your life right now. Connecting those dots has a transportive effect.

Beneficiaries of the event are The Holli Fund and SCAN. For more information on the series and the full list of events, visit instagram.com/summersuppersomm.

Excerpt from:
Vines in Mind - richmondmagazine.com - Richmond magazine

Read More..

Google launches Gemini: The groundbreaking AI project set to outperform ChatGPT – People Matters

Google is embarking on an exhilarating venture with the introduction of Gemini, a groundbreaking project that has the potential to revolutionize the AI industry. Unveiled at this year's Google I/O conference, there is significant anticipation surrounding Gemini's ability to surpass the performance of current AI systems, including OpenAI's ChatGPT.

Google's DeepMind, led by CEO Demis Hassabis, has given birth to Gemini, an innovative AI creation. With an aim to surpass existing AI models like ChatGPT, Gemini operates under the vision of tackling various data and tasks without relying on specialized models. It promises to generate unparalleled content that goes beyond the confines of its training data, marking a significant leap in the field of AI.

Drawing upon the remarkable victory of Google's AlphaGo in 2016, the development strategy for Gemini capitalizes on the techniques that propelled AlphaGo to success. By incorporating AlphaGo's problem-solving capabilities and integrating advanced language processing capabilities, Gemini represents a fusion of these strengths.

The project also embraces reinforcement learning, an iterative approach in which the software continuously endeavors to complete tasks and enhances its performance based on feedback.

With Gemini still in its developmental phase, the anticipated features of this system have already garnered widespread global attention. It is predicted that Gemini will introduce significant transformations in the AI landscape, particularly within the generative AI sector, which is projected to reach a value of 80.16 billion by 2030.

However, it is important to note that Gemini's current capabilities are primarily focused on text processing, unlike GPT-4, which possesses the ability to process images, audio, text, and video. Despite this limitation, Gemini aims to deliver more imaginative and creative responses, aiming to transcend the boundaries of its training data and generate unexpected content.

Previous AI endeavors by Google, such as the chatbot Bard, encountered obstacles that resulted in a factual error during its initial demonstration. This incident had a considerable impact on the market value of Alphabet, Google's parent company. Consequently, the expectations and demands for a flawless launch of Gemini are heightened. It is expected that the introduction of Gemini will undergo meticulous planning to avoid any potential mishaps.

With the ongoing development of Gemini, it emerges as a project of significant interest. The potential success of Gemini holds the power to redefine the AI industry and set unprecedented benchmarks for AI capabilities. Nonetheless, until the final version is released and evaluated in real-world scenarios, the question of whether it will surpass ChatGPT and other AI systems remains unanswered.

Read the original:
Google launches Gemini: The groundbreaking AI project set to outperform ChatGPT - People Matters

Read More..

OpenAI Rival Inflection AI Raises $1.3B to Enhance Its Pi Chatbot – EnterpriseAI

(Ole.CNX/Shutterstock)

Palo Alto-based Inflection AI, an OpenAI competitor, announced it has raised $1.3 billion in a funding round led by Microsoft, Reid Hoffman, Bill Gates, Eric Schmidt, and new investor Nvidia for a total of $1.525 billion raised.

This latest round places the companys valuation around $4 billion, according to a Reuters report.

Inflection AI claims to be building the largest AI cluster in the world along with partners CoreWeave and Nvidia. When completed, the system will be comprised of 22,000 Nvidia H100 Tensor Core GPUs, a release stated.

The deployment of 22,000 NVIDIA H100 GPUs in one cluster is truly unprecedented and will support training and deployment of a new generation of large-scale AI models. Combined, the cluster develops a staggering 22 exaFLOPS in the 16-bit precision mode, and even more if lower precision is utilized, the company said.

The current system contains over 3,500 H100s and recently completed the reference training task of open source benchmark MLPerf in 11 minutes. Inflection AI collaborated with Nvidia and CoreWeave to run the MLPerf tests and fine-tune and optimize the cluster.

The company has developed a large language model, Inflection-1, that enables interaction with Pi, a personal AI chatbot. The company says the new funds will support its continued work of building and supporting Pi. Inflection AI describes Pi as a new class of AI designed to be a kind and supportive companion offering text and voice conversations, friendly advice, and concise information in a natural, flowing style.

Pi, which stands for personal intelligence, is marketed as an AI focused on prioritizing the interests of people both in its functionality and monetization. Imagine your personal AI companion with the single mission of making you happier, healthier, and more productive, wrote Mustafa Suleyman, CEO and co-founder of Inflection AI in a blog post.

Instead of the big tech companies prioritizing advertisers and content creators, Inflection AI is trying a different approach, Suleyman says.

We dont have all the answers, but we are setting out to develop a personal intelligence that really does work for you, thats in your corner, always on your team. Our mission is to firmly align your AI with you, and your interests, above all else. It means designing an AI that helps you articulate your intentions, organize your life and be there for you when you need it, he wrote.

Suleyman, a co-founder of Googles DeepMind, created Inflection AI in 2022 along with LinkedIn co-founder Reid Hoffman and DeepMind alum Karn Simonyan, Inflection AIs chief scientist.

The founders have made Inflection AI a public benefit corporation (PBC), which they say gives them a legal obligation to run the company in a way that balances the financial interests of stockholders, the best interests of people materially affected by our activities, and the promotion of our specific public benefit purpose, which is to develop products and technologies that harness the power of AI to improve human well-being and productivity, whilst respecting individual freedoms, working for the common good and ensuring our products widely benefit current and future generations.

A powerful benefit of the AI revolution is the ability to use natural, conversational language to interact with supercomputers to simplify aspects of our everyday lives, said Jensen Huang, founder and CEO of Nvidia in a release. The world-class team at Inflection AI is helping to lead this groundbreaking work, deploying NVIDIA AI technology to develop, train and deploy massive generative AI models that enable amazing personal digital assistants.

We are very excited to partner with Inflection AI, a pioneering AI company with an outstanding team, to bring the power of supercomputing to cutting edge consumer products, said Michael Intrator, CEO of CoreWeave.

Related

View post:
OpenAI Rival Inflection AI Raises $1.3B to Enhance Its Pi Chatbot - EnterpriseAI

Read More..

5 things about AI you may have missed today: From AI hub, Gemini AI to crime in era of AI and much more – HT Tech

The artificial intelligence race has just got more intense. Google has introduced Gemini, a project by DeepMind, aiming to surpass AI models like ChatGPT; the India Electronics and Semiconductor Association (IESA) has declared Hyderabad as India's AI capital; this AI system, equipped with advanced image recognition, scrutinises waste processing and recycling facilities and AI-Generated images of Zomato's delivery agents go viral on social media - this and more in our daily AI roundup. Let us take a look.

Google introduces Gemini, a project by DeepMind, aiming to surpass AI models like ChatGPT. Gemini's versatility enables it to handle any data or task without specialised models, promising unique content beyond its training data. Building on AlphaGo's success, Gemini combines problem-solving techniques with advanced language processing and reinforcement learning. While limited to text processing, Gemini's creative potential sparks global interest in the rapidly expanding AI industry.

The India Electronics and Semiconductor Association (IESA) declares Hyderabad as India's AI capital, citing excellent leadership and a thriving ecosystem. Hyderabad becomes the prime choice for hosting flagship AI and Machine Learning events. After two years of virtual summits, IESA plans a grand physical event in September. The IESA president commended the city's exponential growth, with ongoing infrastructure development and new global enterprises.

With global solid waste production predicted to soar by 73% to 3.88 billion tonnes by 2050, a UK start-up called Grey Parrot has taken the challenge head-on. Their AI system, equipped with advanced image recognition, scrutinises waste processing and recycling facilities. Over 50 sites in Europe have cameras installed above conveyor belts, enabling real-time analysis of the continuous stream of waste. Greyparrot aims to revolutionise recycling efficiency and address the mounting plastic waste crisis.

A LinkedIn post by Sourabh Dhabhai showcases AI-generated images of Zomato delivery agents enjoying the Mumbai rains. The heartwarming concept reminds us to appreciate their moments of joy amid their work. The post garnered over 6k reactions, with people loving the idea and expressing gratitude for capturing such a sweet moment. AI continues to produce realistic and captivating visuals.

The Union Ministry of Home Affairs will host the "G20 Conference on Crime and Security in the Age of NFTs, AI, and Metaverse" on July 13-14 in Gurugram. In collaboration with other ministries and international bodies, the event aims to address challenges posed by advancing technologies like NFTs, AI, and the Metaverse. Participants from G20 and invitee nations will discuss strategies and implications.

Visit link:
5 things about AI you may have missed today: From AI hub, Gemini AI to crime in era of AI and much more - HT Tech

Read More..

IIT Guwahati rolls out an Online Bachelor of Science (Honours) Degree in Data Science and Artificial Intel – Economic Times

Indian Institute of Technology (IIT) Guwahati is launching an online Bachelor of Science (Hons) degree programme in Data Science and Artificial Intelligence on Coursera, an online learning platform.Anyone after Class XII or its equivalent, with mathematics as a compulsory subject, can apply. Those eligible and registered for JEE Advanced (in any year) will get direct admission, while those without can complete an online course and gain entry based on their performance, according to a release issued by Coursera.This programme teaches students the digital skills they need to thrive in the modern workforce. They graduate knowing how to implement the latest AI and data science techniques in any field, setting them up for success in their careers, said Parameswar K. Iyer, officiating director, IIT Guwahati. Students will receive job placement support from IIT Guwahati and access to Courseras skill-based recruitment platform, Coursera Hiring Solutions, according to the release.

Continued here:

IIT Guwahati rolls out an Online Bachelor of Science (Honours) Degree in Data Science and Artificial Intel - Economic Times

Read More..

In-house automation, analytics tools speed audit processing – GCN.com

Home to one of the countrys hottest housing markets, Travis County, Texasparticularly the city of Austinhas seen the volume of property tax refunds increase by 25% annually since 2018.To keep up and meet requirements to audit the refunds for accuracy throughout the year, the Risk Evaluation and Consulting Division of the countys Auditors Office relies on automation and analytics tools built in-house to perform continuous auditing. REC has reduced the time it takes to process audits of property tax refunds by 91%.

It used to take weeks to analyze the large volumes of property tax refunds, but the model can do it in less than five minutes, said John Montalbo, data scientist for the county. It can detect anomalies, double check for accuracy and write findings to audit standards with incredible efficiency, he added.

Weve gone from 1,000-plus auditor hours per year to [being] at a pace right now for under 40, and we continue to trim that down, REC Manager David Jungerman said. Weve made a lot of progress [in] being able to dedicate folks to more interesting, less mundane work.

Last month, the National Association of Counties, or NACo, recognized RECs work with an Achievement Award for Financial Management.

Even as Travis Countys operating environment and services grew increasingly sophisticated, additional funding for audit compliance was unavailable, according to NACo. Developing innovative, automated auditing techniques allowed auditors to improve their effectiveness and increase their coverage.

The move from a time-consuming, paper-based process has been several years in the making. In 2018, REC began using a dashboard for remote auditing, but the COVID-19 pandemic really showed the office what was possible.

It pushed forward how much more data is being collected during that whole refund process, said John Gomez, senior data scientist at the county. It allowed us to use data to verify when the check was scanned into the system or when the refund application was received and scanned in.

It also enabled auditors to see the metadata so they could determine who looked at and verified an application. Theres a timestamp that gets tied to it recorded and stored, he said.

Since then, the data science team has integrated algorithms into the review process to automate it. Now, human auditors are needed only to review audits that the system calls out as anomalous.

Before the algorithm could be deployed, the data scientists built an extract, transform and load process to collect and organize the data needed for all property tax refunds. Then the countys senior auditor walked them through all the steps she takes and what she looks for in processing the refunds.

We have our algorithms sitting on a virtual machine that will run itself, Montalbo said. Every time that it needs to run, it goes and it gets all the information, does all the tests with which it needs to do, notes exceptions when it finds them, and then starts compiling work documents.

Those documents are put into an email that goes to auditors who spot-check what failed.

Its basically a multi-tab Excel spreadsheet that they get, Jungerman said. We keep one senior [analyst] dedicated to the audit and rotate staff, and basically, they just work the tabs of the spreadsheet if theres any exceptions on there.

Currently, REC is working with the data scientists to automate system-generated receipt testing to streamline audits. Were in the process with 12 county offices right nowand portions of a 13thof looking at all of the system-generated receipts and tracking them to the elected officials bank account and then tracing them to the posting in the enterprise accounting software, Jungerman said. The automation would mean being able to turn around findings to offices within a couple of weeks.

It would also mean processing tens of thousands of receipts every week across all county offices. Currently, receipt testing typically samples only about 80 out of 20,000 receipts, he added.

Automation could be applied to any type of audit, Montalbo said, although the exact mechanisms wont translate seamlessly every time.

We have 50-plus departments [in the county government] and most departments use a different application for their day-to-day business activities, which means different data is being stored for each transaction that is being receipted, Gomez said. So, we have to mine the data for each department to extract the information we need to verify each receipt is recorded correctly and deposited in a timely matter.

Despite the efficiency of automation, Jungerman said that he doesnt foresee any processes running without some form of human interaction. The vision is to automate all of our processes that we can and free standard auditors to just look at exceptions and to look at a whole lot of other areas, he said, adding that you need a human being to verify the potential findings.

Stephanie Kanowitz is a freelance writer based in northern Virginia.

Read the original post:

In-house automation, analytics tools speed audit processing - GCN.com

Read More..

Upcoming online Master’s of Applied Data Science program … – The Daily Tar Heel

Classes for UNC's new Master's of Applied Data Science program, which aims to provide graduate students and working professionals with the ability to advance their knowledge in data science, will begin in January 2024.

To launch this new, primarily online degree, UNC partnered with The Graduate School, the UNC Office of Digital and Lifelong Learning and 2U, an online education company.

To me, the most important aspect of this is that it allows people who are currently in the workforce to have the options of either adding another degree to their resume or to just get some classes, Stanley Ahalt, dean of the UNC School of Data Science and Society, said.

Ahalt said that there are many people who are looking to improve their resume and can benefit financially from developing their data science skills.

Students in the MADS program will have a choice in how they complete this degree. The School of Data Science and Society will offer both live and asynchronous classes.

Arcot Rajasekar will be teaching an introductory course for advanced data science for the new program. Rajasekar is a current UNC professor for the School of Information and Library Science and one of the chief scientists at the Renaissance Computing Institute.

This is for people who are already in the industry and who would like to find new types of tools and techniques and methodologies which are useful for them to deal with their data problems, Rajasekar said.

Rajasekar said that students will gain the proper skills for approaching current and future technology through this program.

"Data science is becoming really important, in a sense, as they call it, 'Data is the new currency,'" he said. "And if you want to deal with data and do large data, what is called big data, you need to have the proper tools to do that."

Kristen Young, the director of communications at the School of Data Science and Society, said that all participants in the MADS program will have the opportunity to participate in an immersion experience that includes staying on campus for two or three days, as well as meeting peers and professors.

This is an experience that we'll provide as an option for online students to have some time on campus and working together in person, Young said.

Ahalt said that there is a high market demand for those with a degree in data science, and that the program would be doing a service for North Carolinians.

According to the U.S. Bureau of Labor Statistics, job growth for data scientists is projected to grow 36 percent between 2021 and 2031. The average employment growth over this time period is five percent.

Students of the program will get to apply their findings to the real world through the MADS capstone projects. For example, Ahalt said the program is considering working with companies in the Triangle to provide students with real-world experiences.

The MADS program not only helps students develop their data science skills, but also equips students with ethical understanding, Rajasekar said. He said that students will also learn how data science can provide avenues for doing good in society.

Those applying to this master's program do not need to have a data science degree, but, Ahalt said the MADS program will require fundamental mathematics, an understanding of programming and a basic working knowledge of some data science modeling.

We're pretty flexible," Ahalt said. "We're going to require some basic skill set coming into the program, but we're trying to make this very accessible."

Applications for the program have been available online since Wednesday, June 21. The deadline for submissions is Tuesday, Nov. 14.

@dailytarheel | university@dailytarheel.com

To get the day's news and headlines in your inbox each morning, sign up for our email newsletters.

More here:

Upcoming online Master's of Applied Data Science program ... - The Daily Tar Heel

Read More..

Data Science: Debunking the Myth of Agile Compatibility – DataDrivenInvestor

A Critical Perspective on Agile Methodologies in Data Science14 min read

Data Science is gaining prominence as a mainstream practice in various industries, leading companies to integrate it into their operations. However, there is a genuine concern that Data Science may be mistakenly categorized as just another software practice, akin to traditional web application development approaches. Over the past years, the Agile hype has spread throughout the technology industry, extending beyond its roots in web development.

Recalling an anecdote, I was once told about Agile being introduced into a legal practice, much to the surprise of the attorneys involved. They found themselves adopting techniques that were completely disconnected from their legal practice, day-to-day work, and actual needs. The resulting negative feedback and disengagement were so overwhelming that they cannot be ignored or understated. The impact was reported to be mentally distressing, almost as if it were an experiment conducted by Dr. Zimbardo[1] himself, the renowned social psychologist.

While coding does play a role in Data Science, it is not the primary activity of a Data Scientist. Unfortunately, this distinction is not widely recognized or understood by individuals outside the field. As organizations grapple with a misunderstanding of what Data Science truly entails, there is an increasing pressure to enforce alignment. In the dynamics of small groups, teams often expand, and IT views Data Science as a logical area for expansion, leading to a perfect storm of misalignment.

To illustrate this point, I once witnessed, with a sense of unease, an Agilist referring to a Data Scientist as a developer and a Notebook with a model as an application. Such remarks highlight the profound misunderstanding of Data Science within the IT industry. It seems that certain factions of the industry adopt a one-size-fits-all mentality, treating a data science study in the same manner as they would approach a web application development project. This approach feels archaic and reminiscent of a bygone era.

Before delving into a detailed analysis of why Agile does not align with Data Science, it is important to understand the origins of Agile and the motivations behind its creation.

The origins of Agile can be traced back to the proclamation of their Manifesto[2] by the Agile Alliance. In general terms, a manifesto is defined as a written statement publicly declaring the intentions, motives, or views of its issuer, as described by Merriam-Webster[3]. Traditionally, manifestos have been associated with movements and schools of thought in social, political, and artistic realms, representing or aspiring to bring about significant qualitative progress for humanity.

Historical examples of manifestos include The Declaration of the Rights of Man and of the Citizen (1789) by the National Assembly of France, The Abolitionist Manifesto (1829) by William Lloyd Garrison, and The Communist Manifesto (1848) by Karl Marx and Friedrich Engels. While individual opinions may differ on the content of these manifestos, they address matters of great importance to humanity. Manifestos are also commonly used to define artistic movements or schools, as exemplified by The Bauhaus Manifesto (1919) by Walter Gropius.

In light of these historical references, it is necessary to express some reservations about labeling a software development methodology, created by software practitioners, as a Manifesto. This usage could be seen as somewhat disrespectful to the likes of Walter Gropius, the National Assembly of France, William Lloyd Garrison, Karl Marx, and Friedrich Engels. It is essential to approach such grandiose associations with caution and scrutiny.

The Agile Manifesto was developed by a group of 17 software practitioners who founded the Agile Alliance. These individuals, including notable names like Kent Beck, Martin Fowler, and Ward Cunningham, are primarily recognized for their involvement in the creation of the manifesto. It is important to note that their expertise lies in coding and related activities, which has formed a consulting industry akin to other domains like coaching and training.

While this association is not inherently problematic, it is worth noting that these authors are not widely acclaimed for groundbreaking software advancements, with the exception of Ward Cunninghams involvement in the creation of the first wiki system. This observation highlights that the Agile Alliance lacks direct connections with industry leaders and innovators.

Recognizing their skill and competence is certainly commendable, and it is not fair or valid to diminish their contributions. However, it does raise questions about the significant impact asserted by the Agile Manifesto without substantial groundbreaking contributions to the field. It prompts us to ponder the underlying motivations and intentions behind the creation of the Agile Manifesto.

Considering who benefits from the Agile Manifesto could help shed light on why it was written and why an Alliance was established. The close association between the members and coaching/training/educational activities raises the question of whether the Agile practice is primarily driven by revenue generation.

While this is a common practice and not inherently wrong, illegal, or unethical, one can infer a conflict of interests that may prevent Agile from being solely focused on your best interests. The assessment of this potential conflict and the alignment of Agile with the broader Consulting industry will depend on your prior experiences with such value-added industries. However, it is widely known that these industries often face criticism due to the lack of accountability and difficulties in measuring performance improvements, especially when they move beyond mere rhetoric.

Recognizing that these reservations are subjective, I will now analyze each claim of the Agile Manifesto individually and evaluate its suitability for the field of Data Science.

These four values are advocated in the Agile Manifesto and form the core of Agile methodology. Lets analyze them in detail:

Data Science is generally considered a lightweight scientific activity. This is because many practitioners in the field primarily apply established methodologies to extract business benefits from data, rather than conducting groundbreaking scientific research. Therefore, the term scientist in data scientist can be seen as more of a vanity term that doesnt fully reflect the pragmatic nature of most practitioners.

However, it is important to note that the Data Science process aims to adhere to scientific methodology in terms of rigor, attention to detail, and employed procedures. It also involves significant mathematical aspects, which are inherently scientific. So, while the intensity of scientific method application may be lower compared to actual scientific research, the underlying aim is still present.

In the context of Data Science, it is difficult to prioritize individuals and interactions over processes and tools. In many cases, the focus is primarily on data, methodologies, and analysis rather than individual interactions. For example, when evaluating the value of a tumor marker, statistical rigor and manufacturing quality are typically more important than the level of individual interaction involved in its development.

Data Science is inherently complex, and in practice, it is often even more challenging to understand and verify compared to regular software development. Jupyter Notebooks have gained popularity because they provide a means of combining inline documentation, including mathematical explanations, with actual code. They resemble traditional scientific research notebooks where authors describe their analysis workflows.

In the context of Data Science, the principle of working software over comprehensive documentation does not align well. An undocumented notebook would be a nightmare scenario, as both the process and outcome of the analysis must be accurately described. In Data Science, comprehensive documentation is just as important as the software itself, if not more so.

In Data Science, the concept of customers is not typically present in the traditional sense. While there may be goals in certain projects, sometimes the work is purely exploratory without specific predefined objectives. Additionally, there are usually no formal contracts in Data Science, as it can be challenging to determine the potential outcomes or directions of a particular study or analysis.

However, it is crucial to clearly specify the specific analysis, outcomes, assumptions, data, and methodology involved in the activity itself. This documentation is typically included as part of the Data Science process, often within the notebook if notebooks are used for analysis. Its worth noting that some Agilists may argue that there are internal customers within a business or organization, as the data generated is intended to be valuable for the overall operation. However, this perspective does not align with the core principle of Data Science.

In summary, while customer collaboration may not be the central guiding principle in Data Science, the clarity and specification of analysis details are essential components of the practice.

In Data Science, traditional plans in the sense of predefined step-by-step procedures are not typically used. Instead, the process often involves formulating hypotheses and testing them or predicting and classifying future events based on past observations. The plan itself becomes an hypothesis to be validated.

However, its important to note that there is usually a script or plan outlining what needs to be done and how to do it. The exploratory nature of Data Science means that outcomes may change and redirect the course of analysis. While this principle of responding to change does not explicitly contradict Data Science, it is not entirely applicable to the field. This principle describes a contradiction that occurs in a different context, such as web application design, where lengthy requirement documents are commonly written and sometimes form part of contractual agreements.

In summary, while Data Science doesnt adhere to traditional plans, there is still a general script or plan in place that can be adjusted based on the evolving insights and outcomes of the analysis.

In addition to the values, the Agilists have principles:

Lets examine them one by one:

(1) Prioritize customer satisfaction by delivering valuable software frequently

In Data Science, the primary focus is on delivering actionable knowledge and insights from data, rather than software. Customer satisfaction is achieved through the quality and impact of the extracted knowledge, rather than the frequency of software deliverables. The value lies in the insights gained, not in the software itself.

(2) Welcome changing requirements, even if they occur late in the project

In the realm of Data Science, the requirements are often based on hypotheses to be tested or predictions to be made. While some flexibility may exist in refining the scope of a project, significant changes in requirements can have far-reaching implications. The iterative nature of Agile may not be as applicable to Data Science, where study cycles are often longer and altering the requirements late in the project can significantly disrupt the study methodology.

(3) Deliver working software frequently, with a preference for shorter timescales

Data Science is not focused on delivering software but rather on extracting meaningful insights from data. The notion of frequent software deliveries is not relevant or feasible in the context of Data Science. The emphasis lies more on the accuracy, validity, and impact of the knowledge extracted, rather than the frequency or timeliness of software releases.

(4) Collaborate with the customer and stakeholders throughout the project

While collaboration with customers and stakeholders is important in any project, it is worth noting that the nature of collaboration in Data Science differs significantly from that in software development. In the initial stages of a Data Science project, interactions with customers and stakeholders play a crucial role in understanding their requirements and objectives. However, once the project moves into the research and study phase, the focus shifts towards extensive data analysis, experimentation, and hypothesis validation, which often occur over longer periods with less frequent interaction.

In Data Science, the emphasis lies on delving deep into the data, applying statistical and mathematical techniques, and extracting valuable insights. This process requires time, careful analysis, and scientific rigor, which may not align with the iterative and rapid delivery approach commonly associated with software development. Therefore, while collaboration remains important, the dynamics of collaboration in Data Science projects differ significantly from those in software development, reflecting the unique nature of the field.

(5) Build projects around motivated individuals and give them the support they need

The idea of building projects around motivated individuals and providing necessary support seems like a self-evident concept applicable to any industry. In the context of Data Science, it is unlikely that professionals would deliberately choose unmotivated individuals or neglect to provide the support required to achieve project objectives.

(6) Measure progress through working software and adjust accordingly

In Data Science, progress is measured by the accuracy, reliability, and impact of the insights generated, rather than by working software. The focus is on refining and improving the analytical models and methodologies based on the data. Adjustments are made to enhance the quality and reliability of the insights, rather than solely based on the functionality of software.

(7) Maintain a sustainable pace of work

While maintaining a sustainable pace of work is important in any field, including Data Science, the nature of Data Science projects may involve extended periods of exploration, experimentation, and analysis. The pace of work may fluctuate depending on the complexity of the data, the methodologies employed, and the depth of insights sought. Striving for a sustainable pace must be balanced with the requirements of the specific project and the need for thorough analysis.

(8) Strive for technical excellence and good design

While technical competence is certainly important in Data Science, the goal is not to pursue technical excellence or intricate design for its own sake. Data Science is focused on utilizing appropriate mathematical tools and methodologies to extract meaningful insights from data. The emphasis lies on the accuracy, validity, and interpretability of the results, rather than striving for technical excellence in the traditional sense.

(9) Keep things simple and focus on what is necessary

The principle of keeping things simple and focusing on what is necessary applies universally to various fields and is not exclusive to Data Science. While simplicity and focus are important, the complexity of Data Science often necessitates specialized techniques and methodologies. The focus is more on deriving actionable knowledge from data rather than oversimplifying or neglecting important aspects of the analysis.

(10) Reflect on your work and continuously improve

The principle of reflection and continuous improvement is valuable in any professional endeavor, including Data Science. However, it is not unique to Data Science and is a widely accepted practice across industries. Professionals in any field are expected to reflect on their work, learn from their experiences, and strive for improvement. Therefore, this principle does not offer specific insights or considerations specific to Data Science.

Summary

In the context of Data Science, it becomes evident that Agile methodologies fall short and are largely irrelevant. The principles put forth by Agile proponents may be seen as nothing more than empty platitudes, failing to address the specific challenges and intricacies of the field. The notion of prioritizing frequent software delivery, embracing changing requirements, and collaborating with stakeholders throughout the project are not only obvious but also fail to recognize the distinct nature of Data Science. Agiles focus on technical excellence and good design disregards the fact that Data Science is more about using the right mathematical tools rather than achieving technical perfection. In truth, Agiles attempt to infiltrate the realm of Data Science can only be described as complete and utter nonsense.

The practical implementation of Agile, particularly in conjunction with the Scrum methodology, often falls short of its intended goals when applied to Data Science. The periodic meetings, known as stand-ups, where team members provide updates, lead to poor engagement and disruption in workflow. The presence of a non-technical or inadequately skilled Scrum master or project manager further compounds the issues, as they normally lack industry-specific knowledge and reduce complex workflows into simplistic task lists. This lack of understanding and accountability creates frustration and hinders the teams progress.

Additionally, the concept of user stories and the emphasis on user-centric requirements do not align well with Data Science, where the focus is more on data, hypotheses, and analysis rather than traditional user-driven needs.

Furthermore, when Agile consulting services are brought in, the emphasis often shifts to methodology and best practices, disconnecting them from the actual business needs and resulting in repetitive and irrelevant discussions. This disconnect and lack of understanding have detrimental effects on team morale and project outcomes, leading to project failures, low quality, massive hidden costs and other negative consecuences.

There is no one-size-fits-all solution for effective project management in Data Science, but based on my experience and observations, the following approaches seem to yield better results:

By embracing these principles, teams can foster a more focused and collaborative environment for Data Science projects.

The Agile Manifesto poses challenges due to its loose definition and sometimes feels akin another Conjoined Triangle of Success[5]. Its values and principles are not universally applicable across industries, and in the realm of Data Science, they often clash with the specific needs and workflows of projects in this field.

As a Data Scientist, it is not uncommon to find yourself pulled into the Agile methodology. However, I encourage you to consider alternatives. Agile is unlikely to serve the best interests of your employer or customers, and it may drain your energy, focus, and time, diverting you from the path of professional growth. Engaging in low-value activities that stray from your core skills can hinder your career.

On a personal and professional level, it is worth considering adjusting your compensation to reflect the challenges posed by following the Agile methodology. The Agile workflow often fosters an environment focused on justifying the methodology itself rather than addressing genuine business needs. Among the various negative aspects, the sense of wasted time can be particularly disheartening. As professionals and human beings, our time is limited, and how we allocate it directly impacts our learning curve and overall fulfillment.

Moreover, Agiles impact on creativity cannot be overlooked. The rigid planning, approvals, timeboxing, and administrative burdens disrupt the very essence of creativity crucial to excelling in Data Science. The prevalence of frequent meetings and administrative tasks stifles the creative process necessary for innovation.

Unfortunately, the prevailing trend indicates that Agile will continue to gain traction. As the world becomes more challenging, we can anticipate an increase in Agile practices.

In conclusion, as professionals, we recognize the importance of navigating the challenges of Agile with resilience as our armor and integrity as our compass. We shall always strive for impactful work, ensuring our actions align with our principles and exemplify professionalism.

I am a seasoned professional with over 20 years of experience in both technical and non-technical roles in technology. I provide contract services to small and medium sized Hedge Funds in AI/Quantitative and Financial Market Data areas.

I live with my family in Denmark in the countryside. If you would like to discuss industry trends, share insights, or explore potential collaborations, I am always happy to connect.

The opinions expressed in this article are solely my own and do not reflect the views or opinions of any past, present, or future employer or customer.

[1] https://en.wikipedia.org/wiki/Philip_Zimbardo

[2] https://agilemanifesto.org/

[3] https://www.merriam-webster.com/dictionary/manifesto

[4] https://de.wikipedia.org/wiki/Politoffizier

[5] https://www.bustle.com/articles/157415-are-the-conjoined-triangles-of-success-real-silicon-valley-mocks-a-famous-business-model

Read the original here:

Data Science: Debunking the Myth of Agile Compatibility - DataDrivenInvestor

Read More..