Page 2,252«..1020..2,2512,2522,2532,254..2,2602,270..»

Clemson data science and analytics master’s program ranked among top in the nation Clemson News – Clemson News

March 2, 2022March 1, 2022

Fortune Education has ranked Clemson Universitys online Master of Science in data science and analytics (DSA) program as one of the best in the country.

Clemsons program, a collaboration between theCollege of Scienceand theWilbur O. and Ann Powers College of Business, ranks 14th in Fortune Educations first-ever onlineranking of data science graduate programs.

There are few programs in the world that two departments develop, said Ellen Breazel, a senior lecturer in the School of Mathematical and Statistical Sciences and the co-coordinator of the program. There are degrees that have both business and statistics, but the management department or the statistics/math department typically teaches them. Clemsons willingness to form a degree program that has equal shares in two departments means our students are getting the experts in both fields.

The demand for data scientists has grown exponentially and theres no sign of it slowing down. TheU.S. Bureau of Labor Statisticsprojects data science-related jobs will grow by 28 percent per year through 2026.

Fortune Educationranking considered selectivity and demand. A programs selectivity score considered the average undergraduate GPA of incoming students, the average number of years of work experience of those students and the programs acceptance rate. The demand score measured the total enrollment size of the program and the number of applicants for the most recent year.

Clemsons first DSA cohort started coursework in 2020.

Many programs speak of cohorts, but the reality is theres a limited difference, said Russ Purvis, professor of management and the DSA program co-director. Our participants are very keen on the importance of such an approach. Some people may think this would not be important for an online program. However, it is quite the contrary. The technical stretches of the program demand students to lean on each other. This and other emotional intelligence skill sets are essential for the workplace and designed to be needed within the program to succeed.

There are currently 51 students in the program.

Seven students graduated last December. An additional 28 students are on track to graduate in 2022.

Craig Fick is one of the programs first graduates. The New Smyrna Beach, Florida, resident worked as a departmental director at a hospital in South Carolina. Some of his job responsibilities aligned with those of a data analyst, such as creating weekly reports and tracking and improving key performance indicators.

I liked doing those tasks, and I knew there had to be easier ways of doing them. Ultimately, this led me to the data science program at Clemson, he said. The program gave me a well-rounded understanding of the data science and analytics world. Coming from a nearly non-technical background, this was very important. The advanced coursework concepts that are covered, I would have had a rather hard time comprehending outside of the program.

The degree comprises 10 courses, five in mathematical and statistical sciences and five in management. There are no required prerequisites, but some background in quantitative reasoning through coursework or work experience is recommended.

Breazel and Purvis said more and more businesses realize the value of big data.

Data-driven decision-making is powerful. Research shows that organizations using business analytics provide strategic planning with information useful in dealing with dynamic environments, Purvis said. Business analytics is also useful when integrated into performance management systems.

For more information, visit the programswebsiteor email msdsa@clemson.edu.

Read more:

Clemson data science and analytics master's program ranked among top in the nation Clemson News - Clemson News

Read More..

Analytics and Data Science News for the Week of March 4; Updates from Alteryx, Grow Inc., Stardog, and More – Solutions Review

The editors at Solutions Review have curated this list of the most noteworthy analytics and data science news items for the week of March 4, 2022. In this weeks roundup, product news from Alteryx and Stardog, and Grow Inc. gets acquired by Epicor.

Keeping tabs on all the most relevant analytics and data science news can be a time-consuming task. As a result, our editorial team aims to provide a summary of the top headlines from the last week, in this space. Solutions Review editors will curate vendor product news, mergers and acquisitions, venture capital funding, talent acquisition, and other noteworthy analytics and data science news items.

The availability of new Alteryx solutions in the cloud means users only need a browser to gain access to insights, with a setup that can be done in minutes. The company is addressing this market need by offering Alteryx Designer Cloud, Alteryx Machine Learning, Alteryx Auto Insights, and Trifacta Data Engineering Cloud in one unified suite, the Alteryx Analytics Cloud.

Read on for more.

Grow offers a no-code, full-stack business intelligence and data visualization tool. The product features data integration capabilities that enable users to connect, store, and blend data from hundreds of data sources. Grow then provides the ability to marry and transform disparate data sources so you can filter, slice, and explore different visualizations. The built-in data explorer defines how you want to navigate data via charts and graphs which are displayed in metrics and dashboards.

Read on for more.

Stardog Designer is a no-code, visual environment for creating knowledge graphs, which helps data and analytics teams to easily apply knowledge graph technology in their work. The tool, along with related innovations completes the companys user experience for data and analytics teams so they can connect to data lakes, visually create semantic data models, and prepare and map source metadata to semantic models.

Read on for more.

Voltron Data is one of the most significant contributors to Arrow. Arrow is a multi-language toolbox for accelerated data interchange and in-memory computing. The Voltron Enterprise Subscription for Arrow is tailored to organizations building and running applications that depend on Arrow. The service offers on-demand assistance from Arrow developers, simplified issue reporting, and direct access to leaders in the project.

Read on for more.

Businesses can now use Cape Privacy to employ encryption-in-use, securely operationalize their most highly classified data, and run predictive machine learning models on encrypted data stored in private clouds or in a third-party data cloud. As a self-service, enterprise-grade platform, Cape Privacy empowers businesses to run as many data models as needed to gain the best possible insights.

Read on for more.

For consideration in future data analytics news roundups, send your announcements to tking@solutionsreview.com.

Tim is Solutions Review's Editorial Director and leads coverage on big data, business intelligence, and data analytics. A 2017 and 2018 Most Influential Business Journalist and 2021 "Who's Who" in data management and data integration, Tim is a recognized influencer and thought leader in enterprise business software. Reach him via tking at solutionsreview dot com.

Follow this link:

Analytics and Data Science News for the Week of March 4; Updates from Alteryx, Grow Inc., Stardog, and More - Solutions Review

Read More..

Winter Collaboratorium to Feature Research from Across the Data Science Institute, Biological Sciences Division, and Pritzker School of Molecular…

Published on Thursday, March 3, 2022

Students file into the Fall 2021 Collaboratorium at the Harper Center. (Photo credit: eClarke Photo)

The Collaboratorium unites University of Chicago students with researchers and faculty who are exploring commercialization opportunities for their work.

The program provides the opportunity for scientists and researchers who want to explore commercialization opportunities to showcase their work and network with students and alumni who may be interested in connecting to pursue further academic study, market research, a business partnership, or participation in an academic competition, such as the New Venture Challenge.

The event will be held in person at the Chicago Booth School of Business Harper Center in Hyde Park with a livestream option available.

>> Register for the Collaboratorium pitch and networking event, here.

The Collaboratorium connects UChicago community members from across campus in ways that they might not otherwise interact, explained Ellen Zatkowski, Polsky Center assistant director and manager of the Collaboratorium. These connections between world-class scientists and talented students foster strong collaborations that have generated enormous impact by bringing cutting-edge technologies to the wider world.

The teams presenting, include:

// Questions? Contact Ellen Zatkowski at ellen.zatkowski@chicagobooth.edu.

Link:

Winter Collaboratorium to Feature Research from Across the Data Science Institute, Biological Sciences Division, and Pritzker School of Molecular...

Read More..

Welcome to the Age of the Engineer-Data Scientist | Transforming Data with Intelligence – TDWI

Welcome to the Age of the Engineer-Data Scientist

The growing enthusiasm for a new hybrid role raises significant questions. We answer them here.

The typical product development/simulation engineering team now enjoys access to a wealth of data that can and should be informing their product design and manufacturing processes. However, finding critical insight within these vast reservoirs of information is another matter. New skill sets are urgently needed. Specifically, engineers must be able to harness artificial intelligence (AI) and machine learning (ML) to support and accelerate better decision-making.

This fundamental shift is illustrated by the emergence of a new hybrid role -- the engineer-data scientist. Whats more, the success of these multi-skilled pioneers will be crucial to the future of the enterprises that are recruiting and training them. Ultimately, engineer-data scientists must shoulder the task of turning the undisputed potential of AI and ML into faster time to market, and they must design more efficient products that perform better for customers and end users.

Its a big ask, and the growing enthusiasm for the new role raises significant questions. Are engineers really the best people to pick up the data science baton? If they are, what skills do they need? On a more practical level, how can they acquire the mindset and capabilities of data scientists? What are the implications for organizations?

Gaining Traction

In engineering and beyond, the data science revolution is gaining traction. A recent PwC survey reports that 86 percent of respondents describe AI as a mainstream technology within their organization. However, in many respects, we have only scratched the surface. A Capgemini report reveals that by deploying AI at scale, automotive OEMs could increase profitability by 16 percent. There is frustration, too. The aforementioned PwC survey also notes that 76 percent of organizations are barely breaking even on AI.

In the search for better return on investment, the creation of the engineer-data scientist is a significant landmark. It reflects growing recognition that solutions should be driven by domain expertise. In other words, the people with granular understanding of the metadata and engineering challenges are the best people to apply the tools that will uncover insight and can thus navigate the best route forward.

Are engineers a good fit for the role? There are convincing arguments in their favor. To start with, although the impact of AI and ML will be revolutionary, it also represents an evolution from what has come before. There are clear parallels with the principles of established engineering techniques such as experiment design, as well as modern simulation and optimization tools. Across every discipline and sector, engineers are comfortable working with simulation, analytical modeling, and statistics.

A Small Step

Of course, the scale and speed at which AI and ML work (and their unprecedented ability to embed continual learning) are revolutionary. At the same time, given their existing capabilities, most engineers will find that embracing data science is a small step rather than a giant leap. By nature, engineers are curious and thrive on solving problems. Moreover, they do not work in the world of pure science. Their focus is on delivering commercially viable solutions. Ultimately, engineers are motivated by a practical desire to build something better. Instinctively, they will be drawn to tools that can help achieve this goal.

Engineers adopting data science are greatly helped by the latest low-code and no-code AI and ML tools. Democratization is hard at work, and the workflows are increasingly familiar and intuitive. However, prospective engineer-data scientists still need to develop new skills. The encouraging news from universities is that data science is an increasingly popular option in engineering courses. Given the urgency of the requirement, we will also see plenty of on-the-job training for more experienced engineers.

Acquiring Skills

Where do I start? is probably the most common question we hear from aspiring engineer-data scientists. The short answer is with algorithms. AI and ML are essentially about matching and applying the right algorithm to the right problem. It is extremely unlikely that an engineer-data scientist will have to take on the job of actually writing these algorithms.

Beyond this, engineers should be encouraged to get involved in projects where they will use AI and ML. From here, we can be confident that their inclination toward hands-on learning is an ideal springboard for a new wave of data science advocates.

This largely organic career path means that engineer-data scientists are always likely to prove an easy fit with their colleagues and organizations in general. Specialist data scientists will remain an important piece of the puzzle, scaling solutions developed by domain experts and building the necessary infrastructures. The difference made by the engineer-data scientist will be seen in design and manufacturing outcomes, not corporate restructures.

If further evidence were needed of the suitability of engineers to take on these new responsibilities, it can be found in a data science sector that is recruiting engineers to fill its own skills gap. Hopefully, the engineer-data scientist role will mitigate the risk of a brain drain. For anyone interested in smarter, more sustainable products, we need our engineers to keep on engineering. We also need them to turn their talents and attention to getting the very best from what data science offers.

About the Author

Brett Chouinard is the chief product and strategy officer at Altair, where he is responsible for the strategy and vision of Altair products, which includes facilitating the development, sales, and delivery of Altairs solutions. You can reach the author on LinkedIn.

Here is the original post:

Welcome to the Age of the Engineer-Data Scientist | Transforming Data with Intelligence - TDWI

Read More..

Women Data Scientists of the World, Unite! – Ms. Magazine

The Women in Data Science Conference was created to ensure women will be represented in the data science field. By 2030, they want 30 percent of all data scientists to be women.The Women in Data Science Conference was created to help women achieve better representation in data science. The conference this year will be held on March 7, 2022the day before International Womens Day. (Courtesy of Women in Data Science Conference)

The Women in Data Science Conference (WiDS) was born of a problem: How can we remove the barriers to success that traditionally bar women from accessing the increasingly critical field of data science?

WiDS co-founder Professor Margot Gerritsen is no stranger to this problem. Gerritsen, who received her Ph.D. in scientific computing and computational mathematics at Stanford University, recalls that as a woman and an international student pursuing a degree in computational science nearly three decades ago, there were few people she felt closely connected toand fewer still who understood the challenges she faced in scientific fields. You cant be what you cant see wasnt yet a slogan, but Gerritsen knew she wanted to help break down the barriers she had faced in the field so that other women would not have to overcome the same obstacles.

Along with co-founders Karen Matthys and Esteban Arcaute, Gerritsen set out to help diversify data science. Their vision of an inclusive future for data science lies at the core of WiDSs mission.

In the current field of data science how data is collected and used as well as who is allowed to collect and use it is extremely limited. Because most data scientists are white men, the kind of data collected and how that data is analyzed often leaves out important groups of people including women, people of color, Indigenous peoples, LGBTQ+ people and more.

Gerritsen points out that these gaps in data science can be quite dangerous since limited perspectives and incomplete and biased sets of data are being used to make decisions that will affect everyone. To Professor Gerritsen, having a diverse group of data scientists at the decision table is vital to creating equitable solutions to the problems we face today.

Diversifying data science also allows people from all groups to access what Professor Gerritsen refers to as the new oil in an evolving economic world: data. Data is a resource that, like oil and gold, gives economic and political power to those that possess it. Gerritsen believes that diversifying data science ensures that people from all backgrounds can access this growing route to powernot just the Elon Musks and Jeff Bezoses of the world.

Because most data scientists are white men, the kind of data collected and how that data is analyzed often leaves out important groups of people.

Diversifying the field, according to Gerritsen, means identifying and removing traditional barriers to entry. WiDS models how to do this through focusing on accessibility. Rather than charging expensive entry fees, WiDS livestreams the conference and provides all WiDS programing for free so those who cannot afford to participate in person are still able to access important information. For Gerritsen, who balanced being a single mother and a full-time worker early in her career, ensuring that women with diverse needs can access WiDS resources is of paramount importance.

Gerritsen also recognizes that a conference in the United States, representing the perspectives of U.S.-based data scientists, could not address the problems women in data science in different regions of the world face. To avoid this potential disconnect, WiDS created local conferences and programming in accordance with local data scientists across the international stage. They ensure that women throughout the world can access resources that speak to their communitys particular needs. At the same time, this model reduces the negative financial and environmental consequences of conference travel.

Of course, many other barriers have been constructed around data science, which Gerritsen is helping to dismantle. WiDS provides mentorships to women in the field who previously have not been able to learn from data scientists that look like them. Gerritsen cites studies showing that for people to feel like they belong in a specific field, about 30 percent of people in that field need to resemble them. This has inspired WiDSs initiative 30 by 30, a project aiming to have 30 percent of people in data science be women by the year 2030.

To Gerritsen, ensuring women can see other women in the field will help them destroy the myth that data science is a field exclusively for men. In constructing the conference and programming around accessibility, WiDS has turned what could have been another expensive and exclusive 20,000-person conference into a network of women working together to find solutions to the problems they face.

WiDSs goal of addressing the unique needs of women of all backgrounds manifests itself in one of the conferences most notable events: the two-month-long Datathon. Every year, WiDS challenges people from all experience levels and fields to work collaboratively with data to solve a problem facing the world.

This year, the Datathon challenged its participants to create solutions to climate change that center on energy efficiency. Working in mixed-gendered teams, participants use their unique backgrounds and experiences to contribute to the efforts of the whole WiDS community.

Gerritsen says Datathon inspires creative and collaborative solutions and creates interest in the field for women of all ages. WiDS provides a model for how to create an international coalition of women in data science, working regionally and nationally to diversify the field and solve some of the most urgent problems of our time.

The Women in Data Science Conference will broadcast live from Stanford University on March 7, 2022, from 8 a.m. to 5 p.m. PTthe day before International Womens Day. Tune into WiDS Worldwide Livestream throughout the day on March 7 to watchkeynotes, tech talks, panel discussions and meet-the-speaker interviews.

Up next:

If you found this articlehelpful,please consider supporting our independent reporting and truth-telling for as little as $5 per month.

More:

Women Data Scientists of the World, Unite! - Ms. Magazine

Read More..

U of T expert on human-centered data science and the problem with the motto ‘move fast and break things’ – University of Toronto

Move fast and break things has become a clich in entrepreneurship and computer science circles.

But Shion Guha, anassistant professor at the University of Toronto's Faculty of Information and a faculty affiliate at the Schwartz Reisman Institute for Technology and Society, says the motto which was once Facebook's internal motto is a bad fit for the technology sector since algorithms are susceptible to biases that can affect human lives.Instead, Guha advocates a human-centered approach to data science that prioritizes the best outcomes for people.

I believe in worlds where data-driven decision-making has positive outcomes, but I don't believe in a world where we do this uncritically, he said. I don't believe in a world where you just throw stuff at the wall and see what sticks, because that hasnt worked out at all.

Guha,the co-author of a new textbook on human-centered data science, spoke to the Schwartz Reisman Institutes Daniel Browne about the need for a more deliberate andcompassionate approach to data science.

Can you tell us about your background?

My academic background is primarily in statistics and machine learning. I graduated with my PhD from Cornell in 2016, and then was an assistant professor at Marquette University for five years before joining the Faculty of Information last year. U of T is one of the first universities in the world to launch an academic program in human-centered data science, so I was nudged to apply.

My co-authors on the book [Human-Centered Data Science: An Introduction, MIT Press March 2022]and I are some of the first people to have talked about the concept of human-centered data science, in a workshop at one of our main conferences in 2016. We decided to write a textbook about the field because we felt there was a missing link between what is taught in the classroom and what happens in practice. In the last few years, the field has talked a lot about algorithmic biases and unforeseen consequences of technology on society. And so, we decided that instead of writing an academic monograph, we wanted to write a practical textbook for students.

What does it mean for data science to be human-centered, and how does this approach differ from other methodologies?

The main idea is to incorporate human-centered design practices into data science to develop human-centered algorithms. Human-centered design is not a new thing;its something that has been talked about a lot in the fields of design, human-computer interaction and so on. But those fields have always been a little divorced from AI, machine learning and data science.

Now, with the advent of this tremendous growth in data science jobs came all of these criticisms around algorithmic bias, which raises the question of whether we are training students properly. Are we teaching them to be cognizant of potential critical issues down the line? Are we teaching them how to examine a system critically? Most computer scientists tend to adopt a very positivist approach. But the fact is that we need multiple approaches, and human-centered data science encourages these practices. Right now, a lot of data science is very model-centered the conversation is always around what model can most accurately predict something. Instead, the conversation should be:What can we do so that people have the best outcomes? Its a slightly different conversation; the values are different.

Human-centered data science starts off by developing a critical understanding of the socio-technical system under investigation. So, whether its Facebook developing a new recommendation system, or the federal government trying to decide on facial recognition policy, understanding the system critically is often the first step. And weve actually failed a generation of computer science and statistics students because we never trained them about any of this. I believe in worlds where data-driven decision-making has positive outcomes, but I don't believe in a world where we do this uncritically. I don't believe in a world where you just throw stuff at the wall and see what sticks, because that hasnt worked out at all.

Next, we engage in a human-centered design process, which can be understood through three different lenses. First, there's theoretical design: the model should be drawn from existing theory what do we know about how people are interacting in a system. For instance, a lot of my work is centered around how algorithms are used to make decisions in child welfare. So, I need to ensure whatever algorithm I develop draws from the best theories about social work and child welfare.

Second, there's something called participatory design, which means inviting all the stakeholders into the process to let them interpret the model. I might not know everything about child welfare, but my models are interpreted by specialists in that area. Participatory design ensures that the people who are affected by the system make the decisions about its interpretation and design.

The third process is called speculative design, which is about thinking outside the box. Let's think about a world where this model doesn't exist, but something else exists. How do we align this model with that world? One of the best ways to describe speculative approaches is the [British TV] series Black Mirror, which depicts technologies and systems that could happen.

Human-centered design practices are about taking these three aspects and incorporating them in the design of algorithms. But we don't stop there, because you cant just put something into society without extensive testing, you need to do longitudinal field evaluation. And Im not talking about six-week evaluations, which are common Im talking about six months to a year before putting something into practice. So, all of this is a more critical and slowed-down design process.

What helps you to collaborate successfully with researchers in other disciplines?

I think one of the major impediments to collaboration between disciplines, or even sub-disciplines, are the different values people have. For instance, in my work in child welfare, the government has a set of values to optimize between spending money and ensuring kids have positive outcomes while the people who work in the system have different values they want each child to have a positive outcome. When I come in as the data scientist, Im trying to make sure the model I build reconciles these values.

My success story has been in working with child welfare services in Wisconsin. When they came to us, I cautioned them that we needed to engage with each other through ongoing conversations to make something successful. We had many stakeholders: researchers in child welfare, department heads, and street-level case workers. I brought them together many times to figure out how to reconcile their values, and that was one of the hardest things that I ever did, because people talk about their objectives, but don't often talk about their values. It's a hard thing to say, OK, this is what I really believe how the system should work.

We conducted workshops for about a year to understand what they needed, and what we eventually realized was that they were not interested in building an algorithm that predicted risk-based probabilities, they were interested in something else: how to make sense of narratives, such as how to describe the story of a child in the system.

If a new child comes into the system, how can we look back and consider how this child displays the same features as other historical case studies? What positive outcomes can we draw upon to ensure this new child gets the services they need? It's a very different and holistic process its not a number, it's not a classification model.

If I had just been given some data, I would have developed a risk-based system that would have ultimately yielded poor outcomes. But because we engaged in that difficult community building process, we figured out that what they really wanted was not what they told me they wanted. And this was because of a value mismatch.

Similarly, when I go to machine learning conferences, theres a different kind of value mismatch. People are more interested in discussing the theoretical underpinnings of models. I am interested in that, but Im also interested in telling the story of child welfare, Im interested in pushing that boundary. But a lot of my colleagues are not interested in that their part of academia values optimizing quantitative models, which is fine, but then you can't claim you're doing all these big things for society if that's really what your values are.

It's interesting to note how much initial effort is required, involving a lot of development that many wouldn't necessarily consider as part of system design.

You know, the worst slogan that Ive ever heard in the technology sector, even though people seem to really like it for some reason, is move fast and break things. Maybe for product recommendations that's fine, but you don't want to do that if you've got the lives of people on the line. You can't do that. I really think we need to slow down and be critical about these things. That doesn't mean that we don't build data-driven models. It means that we do them thoughtfully, and we recognize the various risks and potential issues down the line, and how to deal with it. Not everything can be dealt with quantitatively.

Issues around algorithmic fairness have become very popular and are the hottest field of machine learning right now. The problem is that we look at this from a very positivist, quantitative perspective, by seeking to make algorithms that are mathematically fair, so different minority groups do not have disproportionate outcomes. Well, you can prove a theorem saying that and put it into practice, but heres the problem: models are not used in isolation. If you take that model and put it where people are biased, when biased people interact with unbiased, mathematically fair algorithms it will make the algorithms also biased.

Human-AI interaction is really important. We can't pretend our systems are used in isolation. Most problems happen because the algorithmic decision-making process itself is poorly understood, and how people make a particular decision from the output of an AI system is something we don't yet understand well. This creates a lot of issues, yet the field of machine learning doesn't value that. The field values mathematical solutions, except it's a solution only if you view it in the context of a reductionist framework. It has nothing to do with reality.

What are some of the challenges around the use of algorithmic decision-making?

My co-authors and I identify three key dimensions of algorithmic decision-making. One dimension is that decisions are mediated by the specific bureaucratic laws, policies, and regulations that are inherent to that system. So, there are certain things you can do, and cant do, that are mandated by law. The second dimension is very important, we call it human discretion. For example, police may see a minor offense like jaywalking but choose to selectively ignore it because they are focused on more significant crimes. So, while the law itself is rigid, inside the confines of the law there is discretion.

The same thing happens with algorithmically mediated systems, where an algorithm gives an output, but a person might choose to ignore it. A case worker might know more about a factor that the algorithm failed to pick up on. This works the other way too, where a person might be unsure and go along with an algorithmic decision because they trust the system. So, theres a spectrum of discretion.

The third aspect is algorithmic literacy. How do people make decisions from numbers? Every system gives a separate visualization or output, and an average social worker on the ground might not have the training to interpret that data. What kinds of training are we going to give people who will implement these decisions?

Now, when we take these three components together, these are the main dimensions of how people make decisions from algorithms. Our group was the first group to unpack this in the case of public services, and it has major implications for AI systems going forward. For instance, how you set up the system affects what kinds of opportunities the user has for exercising discretion. Can everyone override it? Can supervisors override it? How do we look at agreements and disagreements and keep a record of that? If I have a lot of experience and think that the algorithms decision is wrong, I might disagree. However, I might also be afraid that if I don't agree, my supervisor will punish me.

Studying the algorithmic decision-making process has been crucial for us in setting up the next series of problems and research questions. One of the things that Im very interested in is changes in policy. For example, my work in Wisconsin was utilized to make changes that had positive outcomes. But a critical drawback is that I haven't engaged with legal scholars or the family court system.

One of the things I like about SRI is it that brings together legal scholars and data scientists, and Im interested in collaborating with legal scholars to think about how to write AI legislation that will affect algorithmic decision-making processes. I think it demands a radical rethinking of how laws are drafted. I don't think we can engage in the same process anymore; we need to think beyond that and engage in some speculative design.

What is the most important thing that people need to know about data science today, and what are the challenges that lie ahead for the discipline?

Obviously, Im very invested in human-centered data science. I really think this process works well, and since U of T began its program, the field has expanded to other universities and is gaining momentum. I really want to bring this to the education of our professional data science students those who are going to immediately go out into industry and start applying these principles.

Broadly, the challenges for the discipline are the problems I've alluded to, and human-centered data science responds to these issues. We should not be moving fast, we should not be breaking things not when it comes to making decisions about people. It doesn't have to be high stakes, like child welfare. You can imagine something like Facebook or Twitter algorithms where ostensibly you're doing recommendation systems, but that really has ramifications for democracy. There are lots of small things that have major unintended consequences down the line, even something like algorithms in the classroom to predict whether a child is doing well or not.

The other main challenge is this value mismatch problem I described. We need to teach our next generation of students to be more compassionate, to encourage them to think from other perspectives, and to center other people's values and opinions without centering their own. So how do we get better? Again, human-centered design has worked very well in other areas, and we can learn what worked well and apply it here. Why should we pretend that we have nothing to learn from other areas?

See the original post here:

U of T expert on human-centered data science and the problem with the motto 'move fast and break things' - University of Toronto

Read More..

Nontraditional Student Will Be a Double Hoo, 33 Years Apart – UVA Today

When Maureen OShea heard about UVAs School of Data Science, the youngest of her six children was about to graduate from high school.

OShea, who spent the last two decades living in Charlottesville, decided to attend the Masters in Data Science open house. After listening to faculty members speak, she left feeling inspired and thought, I think I can do this.

In 2021, OShea, who will graduate this spring, began her first semester in the School of Data Science. She took courses that interested her and helped her find a tightknit community. These students are from a variety of fields and ages, something OShea cites as a benefit of the program.

They offer such a broad variety of classes, and its very interdisciplinary, so people are coming from all different backgrounds, she said. Our cohort is only about 50 students, so everybodys familiar, and you can get to know the professor and the teaching assistants.

The School of Data Science offers courses in data ethics, preparing students with decision-making tools. OShea noted that these skills are essential for becoming a data scientist.

In our first semester, we took a class in data ethics, which I found to be really important. As data scientists, we have to take ethics very seriously and question what people are asking you to do with the data, because the way you tell the story with the data can change decision-making.

This isnt OSheas first time as a student on Grounds. After receiving her undergraduate degree from Loyola University in Baltimore, she came to UVA in 1986 for a masters in material sciences and engineering.

The masters program was incredible and also where I met my husband, who was getting his masters in mechanical engineering, she said.

After receiving her UVA degree, she worked at IBM in their failure analysis department. She spent five years there before opting out of the work force to be with her children.

I really enjoyed IBM, and I was fascinated with failure analysis because Im very curious. I always like to know why? and dig into the nitty gritty of things, OShea said. We eventually decided that Id like to stay home and raise our kids. And then, you know, six kids later, I was still home, and I wouldnt trade it for the world.

Now back at UVA, 33 years later, technology has changed the classroom. OShea has noticed a clear difference in how professors teach using technology and how students interact with the class material.

By far, the biggest change between my two UVA experiences has been the advancement in technology. In the 1980s, we didnt have laptops and went to the computer lab to type our papers, she said. I admire and am amazed at how students today are able to take notes on their computers or iPads during class lectures, even though I still take notes the old-school way, by hand.

OShea said a key component of her success in returning to UVA was her preparation prior to the program. She took courses at Piedmont Virginia Community College before starting in the School of Data Science.

I went to PVCC for a semester and brushed up on some skills before joining the Data Science program and I think its huge to familiarize yourself with some material so youre able to hit the ground running, she said. I was sitting in a calculus class at PVCC, and the girl I was next to was someone who graduated with my daughter, which was really funny.

After her second Final Exercises walk down the Lawn, OShea hopes to continue volunteer work with organizations in Charlottesville and East Africa. With new skills from the School of Data Science, she can provide these organizations with critical assistance.

During my time at home, Ive been involved in a number of volunteer organizations. Ive worked with The Haven, a homeless day shelter, to help their guests file taxes. I would like to go back and use the skills Ive learned to help places like The Haven more formally, she said.

Im also involved in an investment angel network that supports entrepreneurs and private sector development in East Africa. Through impact investing, we fund small/medium-sized East African companies with investments and loans. I have a soft spot for the women-owned and educational enterprises.

As OShea prepares to graduate, she is leaving UVA with a desire to help others and gratitude for her time on Grounds.

Just maybe Ill come back for a third graduate degree, OShea said.

Continue reading here:

Nontraditional Student Will Be a Double Hoo, 33 Years Apart - UVA Today

Read More..

10m Irish project to focus on treating ALS with data science and AI – Siliconrepublic.com

Precision ALS will bring together clinicians, data scientists and AI experts to find new ways to treat motor neurone disease.

A new 10m Irish project is looking to develop new and innovative treatments for patients with motor neurone disease by bringing together medical research, data science and artificial intelligence.

Launched today (1 March) at Trinity College Dublin (TCD), Precision ALS will build tools to enable clinical trials based on precision medicine, where treatments are personalised for individual patients.

The project will be led by two Science Foundation Ireland (SFI) research centres: the Adapt centre for AI-driven digital content technology and the FutureNeuro research centre for neurological diseases.

It will bring together clinical scientists, data scientists and AI experts to collaborate on data-driven prediction models for progression of the neuromuscular disease and data analysis that will help develop treatments.

Prof Orla Hardiman, who is the director of Precision ALS and a professor of neurology at TCD, said that there is an increasing recognition of the need for precision medicine in developing drugs for motor neurone disease, or ALS, which only affects humans.

ALS is also a heterogeneous disease, she explained, which means that it has many different causes and patterns of progression, and a large amount of data is required to understand these differences.

Using big data analyses, Precision ALS will provide an in-depth understanding of the factors that drive heterogeneity, and in doing so will for the first time allow us to target new and innovative treatments to specific patient subgroups, Hardiman said.

The research is supported by the Irish Government through an SFI investment of 5m, which will be matched by an additional 5m from industry partners.

Speaking at the project launch, Tnaiste and Minister for Enterprise, Trade and Employment Leo Varadkar, TD, said that Precision ALS will combine the best of our technologies, the best of our ideas, and the best of our medical research to change the lives of patients living with the disease.

It will develop tools that facilitate clinical trials based on precision medicine and has the potential to produce benefits for other rare conditions and diseases, supporting job creation and reducing drug costs, he added.

The project will also provide an interactive platform for clinical research in ALS across Europe, which will use AI to analyse large amounts of data gathered at scale and in a timely and cost-effective manner across multiple international sites.

Prof Vinny Wade, director of the SFI Adapt centre, said that Precision ALS brings together a perfect mix of data and technology research skills to trailblaze discoveries in tackling these devastating diseases.

Wade believes that the centres experience in researching data sets for immediate interrogation using AI will help identify contributing factors and help discover changes linked to ALS.

Unlocking this data in an ethical way is the key to achieving the research mission and realising true precision medicine. This pioneering work will lead to transformational change for patients with a ripple effect that will positively impact society, he added.

The Ireland-based researchers will work in partnership with TRICALS, an independent consortium of leading ALS experts, patients and patient advocacy groups across Europe. Companies partnering with Precision ALS include Biogen, Novartis, Takeda, IQVIA, Roche and Accenture.

Dont miss out on the knowledge you need to succeed. Sign up for theDaily Brief, Silicon Republics digest of need-to-know sci-tech news.

See original here:

10m Irish project to focus on treating ALS with data science and AI - Siliconrepublic.com

Read More..

For Six Sigma Black Belts: It’s Time To Break Fresh Ground With Sustainable Process Performance – Manufacturing.net

When manufacturing processes are no longer effective, what do you do?

The upheaval suffered by manufacturers since the onset of the COVID-19 pandemic is forcing companies to rethink their processes. Its not surprising that Six Sigma thinkers, as process improvement experts, are high on the go-to list for help.

But for Six Sigma experts, its also time to rethink.

To see why lets take an example: Back in 2012, our process improvement team completed a DMAIC (design, measure, analyze, improve, control) project to improve the process performance of a plants production line. Their recommendations were operationalized and they concluded the project by installing a control system.

The purpose of the control system was to ensure continued performance of the new process at the plant. Any dip would be caught and rectified. This sounds good, but move forward now to 2022: The control system still validates performance based on the data that was collected 10 years ago.

Its highly likely that the data used in 2012 upon which the process was set up is no longer relevant. So why is new data never used by the control system?

In short, becauseDMAIC does not iterate between phases.

The DMAIC control phase is isolated from the analysis phase. In our DMAIC project, we collect the data, define the problem, run our analysis, operationalize our recommendation and define a control system to ensure process performance. But the control phase never revisits the analysis. And our analysis is never re-run based on new data.

We need to merge the analysis and control phases to enable a continuous loop: Analyze Operationalize Control Analyze Operationalize Control.

Over in the data science world, data scientists take an iterative approach, known as thedata science life cycle. They are collecting the data (in real time) and learning from the data to then improve, describe and predict an outcome/produce an analysis. This outcome is what the data scientists call their model. To ensure that the model continues to perform well when operationalized, it is continually monitored and, whenever necessary, updated. Should there be a dip in performance, they go back to retrain (update) their model on new data, i.e., reanalyze and re-operationalize.

Manufacturers are spending a lot of time, money and effort to improve their processes. In DMAIC, we also need to see how we can improve our process. To be truly sustainable, we need to merge the analysis and control phases and create an iterative cycle. We need to add data science.

As we monitor performance of our production line process, new data is flowing into our database on a daily basis from the sensors fitted to the machines. Typical statistical packages are unable to handle todays high volume of data at high frequency; the effort to collect data also from multiple sources is huge.

Using a data science tool, we have the means to not only easily process huge volumes of data but also do so in real time. We can easily read in the 5 million datasets produced by our production line. Our data science solution literally learns from the data, evaluates the data and produces a model. But it doesnt stop there.

Our data science tool lets us go back to the beginning of the cycle and optimize our process based on the new data.

Now when new people come into the process, e.g., a new supplier is needed for a certain part and new lines are developed, we can feed this new data into our analysis. We can reuse our model: It learns from the new data and produces an optimized result.

A single workflow, for example, enables us to read in the 5 million datasets produced by the machine each day, evaluate this data, show us a visual pre-scan and then run the data through three different models before providing a comparison to tell us which model is performing best. It only takes seconds to run.

The process is now immediately responsive to any changing circumstances reflected in the data. With our DMAIC tool, we would have needed to start an entirely new project to solve the issue.

Fig. 1. KNIME workflow for data preparation, dataset evaluation, visual pre-scan of the data, and model building.Six Sigma

At any stage in the project, we can go into our analysis and check the output of different stages. We can inject our knowledge as process experts, for example, examine the correlation and numeric outliers to get a sense of the quality of the data and tweak as needed. We can use the pre-scan to interactively zoom in to inspect a group of figures in more detail.

If we see that something is wrong, we can immediately go back a step, make an adjustment and rerun the workflow.

Six Sigma

Fig. 2. Two pre-scan visualizations showing sensor data in an interactive parallel coordinates plot and scatter matrix.Six Sigma

In a DMAIC project, we tend to define a single hypothesis, using regression analysis to measure whether results align with what we are expecting. But are we comparing our regression analysis with any other model type? Probably not.

With our workflow, however, we can not only regularly evaluate how our model is performing but also set up multiple models and evaluate how they are all performing.

In our example, a visualized comparison shows us the quality of our three models. The results: Decision tree 0.91-very high, Naive Bayes 0.73-also good, Logistic Regression 0.74 show us that although our regression p value is OK, the decision tree is performing better. In typical Six Sigma tools, analysis techniques such as decision trees or Naive Bayes are not available options.

We can also decide to run each model based on 10 different test and training sets and it takes only a second. It provides us with failure rates and visualization for each scenario.

With our data science solution, we can regularly evaluate our process, it is able to respond quickly to changes in the data and we can compare based on a range of models if performances are changing, check why and deploy the best process.

We can even automate this entire cycle.

By enabling the control system to be automatically monitored, evaluated and (re)deployed, we ensure not only that it gets done reliably but also produces much more accurate results. When you tell a machine to control a process, it just does it. And keeps on doing it.

---

Andreas Riess is an expert in Six Sigma and quality engineering. As a certified trainer and supporter, with 15 years atMTS Consulting Partner, he is recognized on both national and international levels and offers consulting services throughout the preparation and ramp-up of new production lines based on a Six Sigma-supported structured approach. His work is backed by 25 years of experience at a global automotive technology company that supplies systems for passenger cars, commercial vehicles and industrial technology. Andreas earned his engineering degree from The University of Applied Sciences Wrzburg-Schweinfurt.

Heather Fyson is a content writer in data science forKNIME, a data analytics software company bridging the worlds of dashboards and advanced analytics through an intuitive interface, appropriate for anybody working with data. KNIME is distinct in its open approach, which ensures easy adoption and future-proof access to new technologies.

Read the original post:

For Six Sigma Black Belts: It's Time To Break Fresh Ground With Sustainable Process Performance - Manufacturing.net

Read More..

Alation Celebrates International Women’s Day with 2nd Annual Women Leading Data Intelligence Event – Business Wire

REDWOOD CITY, Calif.--(BUSINESS WIRE)--Alation Inc., the leader in enterprise data intelligence solutions, today announced that it will host its second annual Women Leading Data Intelligence event. The event will take place on International Womens Day, Tuesday, March 8, 2022 at 8 a.m. PT, and feature a panel of exceptional individuals from Fidelity Investments, Snowflake, and WEX, that are driving data culture in their organizations. To register for the digital event, visit https://events.alation.com/IWD2022.

The women that power data and technology organizations represent a massive talent pool often overlooked or undersupported in male-dominated industries. To support diversity in data careers, Alation created Women Leading Data Intelligence, which provides a forum for networking, education, and career development for women in analytics, data science, machine learning, and artificial intelligence.

The event panel, moderated by Jen Wu, Senior Director of Product Management at Alation, will feature Jennifer Belissent, Principal Data Strategist at Snowflake; Meredith Frinsko, Director, Data Strategy & Governance at WEX; and Susanne O'Neill, Workplace Investing - Data Governance COE Leader at Fidelity Investments. Julie Smith, Director of Data & Analytics at Alation and Tracy Eiler, Alation Chief Marketing Officer and founding member of Women in Revenue, will host an intimate fireside chat following the panel.

Women are a growing force in data and technology, and we are proud to showcase these leaders driving data culture and change within their organizations, said Eiler. The impact of women in our workforce is something that should be recognized throughout the year and we are excited to bring together some of the most influential voices in our industry to highlight their innovative contributions.

The most successful and productive organizations have a diverse employee-base, and gender is a key aspect of diversity. At Alation, 30% of employees are women who lent their work ethic and tenacity to the companys achievements this past year. Alations executive leadership team is nearly 35% women, most recently adding Langley Eide as Chief Financial Officer. The company welcomed Preeti Rathi, General Partner at Icon Ventures, to its board of directors in October 2021. According to the Top Companies for Women Technologists Report that surveyed more than half a million technologists, women represent only 26.7% of the technical workforce on average.

In May 2021, Alation was named one of Inc. Magazines Best Workplaces. To be considered for the list, Alation employees participated in the magazines survey, and 79% of women cited that they saw professional growth and career development opportunities for themselves at Alation. In the last year, Eiler founded a Women at Alation group where women Alationauts participate in quarterly events to discuss topics such as the wage gap, career growth, and mentorship.

To continue supporting women in data and technology, Alation is a proud sponsor and supporter of Women in Data and Women in Revenue for the second consecutive year. Women in Data was founded to increase diversity in data careers, and provide awareness, education, and advancement to women in technology, specifically analytics, data science, machine learning, and AI. Women in Revenue was founded in 2018 by 12 revenue-driven women. The organizations mission is to achieve workplace equity through networking opportunities, mentorship, and education.

Learn More:

About Alation

Alation is the leader in enterprise data intelligence solutions including data search & discovery, data governance, data stewardship, analytics, and digital transformation. Alations initial offering dominates the data catalog market. Thanks to its powerful Behavioral Analysis Engine, inbuilt collaboration capabilities, and open interfaces, Alation combines machine learning with human insight to successfully tackle even the most demanding challenges in data and metadata management. More than 330 enterprises drive data culture, improve decision making, and realize business outcomes with Alation including AbbVie, American Family Insurance, Cisco, Exelon, Fifth Third Bank, Finnair, Munich Re, NASDAQ, New Balance, Parexel, Pfizer, US Foods, and Vistaprint. Headquartered in Silicon Valley, Alation was named to Inc. Magazines Best Workplaces list and is backed by leading venture capitalists including Blackstone, Costanoa, Data Collective, Dell Technologies, Icon, ISAI Cap, Riverwood, Salesforce, Sanabil, Sapphire, and Snowflake Ventures. For more information, visit alation.com.

Read this article:

Alation Celebrates International Women's Day with 2nd Annual Women Leading Data Intelligence Event - Business Wire

Read More..