Page 759«..1020..758759760761..770780..»

Need to manage your Google Drive data? Google was no help, but this tool solved my storage mess – ZDNet

Andriy Onufriyenko/Getty Images

One of the oddest things about Google Drive cloud storage is that it doesn't provide a mechanism for you to be able to tell how much storage a given folder takes up. Google is quick with how much overall storage you use, because that's how they bill you. But if you want to know which folders take up how much space, there's no cloud interface for that.

When you ask Google's support team, their answer is this: Set up syncing and then use your local OS to tell you how much data you have in the synced directory. This is all well and good when you have 10GB in synced cloud storage, or even a terabyte or two.

Also:The other shoe finally dropped on my Google Enterprise cloud storage plan

But when you have an enterprise plan, and you have 58TB of cloud storage in use, it's not like you're going to download your data to a local computer just to find out which folders are taking up a lot of space.

I faced this challenge recently when Google capped my unlimited storage. I'm still smarting a bit about how unlimited always has a limit. Fortunately, after a lot of begging, Google did give me a bit more storage, so I was no longer suddenly at capacity. But the writing is on the wall. I need to prune my online storage usage.

But even then, I still face the problem I described above -- how much storage is each of my folders taking up? I can't prune if I don't know what I'm pruning.

Also: How to easily manage your Google Workspace storage

Fortunately, while Google doesn't offer a solution, a developer named Brett Batie in Traverse City, MI has come up with a solution. It's called Filerev and it does just what you would have expected Google Drive to do. It answers questions and helps manage vast amounts of data.

When we last encountered this issue, I was trying to optimize my storage utilization, so that it didn't grow beyond its new 75TB bounds.

My remote Google Drive storage is divided into two main folders: Cloud Backup, and Cloud Sync.

The Cloud Backup folder holds incremental backups of my servers. With this setup, I can keep all files, including those that have been deleted. That way, I can execute a restore and grab the state of a folder or file on any given date.

Also:Why that cheap 'lifetime cloud storage' deal might cost more than you bargained for

Cloud Sync is a 1-for-1 sync to the server folders, as they are right now. This folder allows me to get to a live file remotely, without having to go through the very long and convoluted process required by a restore.

The Cloud Sync folder has come to my rescue numerous times when traveling, such as when I had to evacuate Florida when the hurricanes struck.

The picture above shows something quite odd. I did a backup early this morning. And yet, the "Last modified" date shown is from 2018. When I drill down to a file I know I modified a day or so ago, it does show the current date. So, why is that happening? It's so frustrating. If that folder has changed, it should contain the date of the latest change. But no, it's just not that easy.

As for my ongoing challenge with finding the size of folders, you would think that by digging into the menu for a folder, as I've done here, the File Information > Details menu item would show additional details, such as how big a folder might be. But all it does is open Details pane, which, unamusingly, does not contain the details we need.

As for the Details pane (pain?), it doesn't contain any data about size at all. It does include the correct date for the last time a folder was opened, but somehow "modified" still shows 2018, even though I modified it yesterday by adding some files and folders.

So, how on Earth can I determine how much storage I am using in each of my folders? We're talking about an enterprise account with 58TB of storage. And yet, there's no visibility into storage utilization. It's not just bizarre, it's deeply inconvenient.

Batie's Filerev software seriously comes to the rescue in this situation. Filerev is available in a variety of different plans. I won't be doing a lot of scans, but the scan I did reveals that I have 4,875,461 files (a number I never would have known without Filerev), so I definitely need the Premium plan.

Batie was kind enough to give me a free account. He also went above and beyond, and answered my questions while I was struggling to understand my file utilization. He incorporated my feedback and added a number of useful features, including the drill-down graphics, which I basically begged him for.

Also:Google's Duet AI for Workspace can create presentations, write emails, and attend meetings for you

Filerev ascertains your storage structure by doing a file scan. For me, that took a few days. It's also a bit of a worry, because you have to open up access to your files to Filerev. But while I am a bit concerned by that security issue, Batie says that the tool retrieves mostly metadata and not file contents. Given the choice between losing access to all my data due to Google's limits on unlimited usage, and having this tool explore my metadata, opening up to Filerev is the lesser of two evils.

Here's the most interesting detail. I can see how my Cloud Sync data compares to Cloud Backup:

Clicking the More button gives me actual storage details. The big shock was that the Cloud Backup folder uses considerably less storage than the Cloud Sync folder. Until I saw that graphic, I had been convinced it was the other way around.

You can click into either folder. Clicking into the Cloud Sync folder tells a very interesting story:

Two of my top three storage usages make sense. The folder Backup Share is a mirror of the local machine backups on my local area network, which are different to the overall server backups that are in the main Cloud Backup folder.

And the Studio Share folder, at 7.42TB, makes total sense. That's where I keep all my video work-in-progress for the various videos on my YouTube channel.

Also:The best cloud storage services of 2023: Expert picks

But it's Liberty Tank, at 11.1TB, that's most instructive. This was a server I previously backed up, but it's long out of date. All the files on Liberty Tank were moved to the current server and are being backed up separately. So the sync with Liberty Tank, all 11.1TB of it, is unnecessary storage utilization.

Just nuking that, alone, will move my storage utilization from a stressful 58TB down to a more calming 47TB.

And there's no way I would have known that without Filerev.

Filerev has quite a few additional features, including a lot of customization. But one feature I'd like to point out is the File Groups summary:

Although the closeness of the red-like colors make it a bit hard to tell, my videos do not take up the most storage. That honor goes to the category of Other.

Drilling into Other shows this Octet-Stream category:

As it turns out, that storage category accounts for all my virtual machines. And yes, it makes total sense that I have 36TB of virtual machines (VMs). I do a lot with VMs. More to the point, I once did a tremendous amount with VMs. For many years, I was running multiple virtual server racks on my computers that I used as part of my development process.

It won't be easy to trim down those VM syncs, but it's definitely something I can look at, because I know a bunch of those VMs are very obsolete.

I made a few interesting discoveries. My Cloud Backup share isn't the villain I thought it was. That honor goes to the Liberty Backups folder. Second, the biggest category of storage is my storage of old VMs. That's a homework assignment I need to follow up on to clear out more space.

But also, this is a story of individual developers and the contribution they make to the industry. Filerev does what Google Drive should do. There's really no excuse for an enterprise-level cloud storage service to lack basic directory analytics.

But this one dude in Traverse City, Brett Batie, stepped up and built a solution and a business. Not only did he provide a useful software solution, he also collaborated with me on some interesting value-adding features to Filerev, which helped me solve my problem and I'm sure will help others.

Batie's efforts show the best of the indy developer phenomenon, and I very much appreciate his help. Give Filerev a look. It might help solve a critical storage situation for you, as well.

You can follow my day-to-day project updates on social media. Be sure to subscribe to my weekly update newsletter on Substack, and follow me on Twitter at @DavidGewirtz, on Facebook at Facebook.com/DavidGewirtz, on Instagram at Instagram.com/DavidGewirtz, and on YouTube at YouTube.com/DavidGewirtzTV.

Continue reading here:
Need to manage your Google Drive data? Google was no help, but this tool solved my storage mess - ZDNet

Read More..

More than meets the eye: The full value of cloud archiving – NewscastStudio

Subscribe to NewscastStudio for the latest news, project case studies and product announcements in broadcast technology, creative design and engineering delivered to your inbox.

There is a preconception among media and entertainment (M&E) companies that the cloud can be viewed simply as a convenient storage hub. This view that cloud technology equates to storage has caused some inertia for adoption across the M&E industry as budgets tighten. In reality, storage is only one facet of the clouds capabilities, and broadcasters are quickly discovering its potential for actively generating revenue and driving business operations.

Cloud archiving is a tool which offers a wider range of business benefits than many people realize there is, in fact, far more to be leveraged through cloud archiving than gains to storage capacity. Broadcasters are beginning to recognize the potential of cloud archiving as an engine for driving business operation efficiencies and unlocking new avenues for revenue growth.

M&E companies face the daunting prospect of a highly competitive market that demands uninterrupted delivery of quality content. This is a tall order, and tools that can be leveraged to meet these demands are being sought with growing urgency as the global appetite for content increases. Operational efficiencies must be continually refined and improved in order to keep pace in todays crowded market. To plan forward and build long-term business agility, the choice of tools must be highly flexible. The adaptable nature of cloud technology enables practically unlimited scalability and this is an essential precondition for broadcasters to thrive in todays rapidly expanding market.

One of the clouds most powerful tools for driving revenue lies in its archiving capabilities, which can be used to turn archive material into newly profitable content thats ready for broadcasting. For companies with extensive archives, the cloud offers a way to realize the full value of what theyve been sitting on; its a recipe for turning lead into gold.

Conversely, building and maintaining physical facilities for local archiving is an ongoing cost that only the largest industry players can afford let alone justify as smaller competitors can make better use of their existing content without the burden of manually storing and managing their archives on site.

One of the essential ways that the cloud makes archive content usable is through metadata enhancement. Metadata enhancement is an automated process that applies metadata to archive content tagging and categorizing the archive material by set criteria (which can be adapted to suit the individual use cases for each business). This means that any content can be sorted and retrieved instantaneously to suit the business needs of operators at any given time a great asset to keep viewers watching after the expensively produced show they tuned in to see has finished. With metadata keeping archive material accessible by an automated process, the gains to business agility and the benefits for the production workforce make metadata one of the most powerful tools that the cloud has to offer.

To meet the round-the-clock requirements of video consumption in the digital age, the tools used by video production teams must be optimized for flexibility. For creative collaboration, cloud storage is the closest thing to a frictionless solution. Its the tool that provides a basis for creative teams to thrive equally whether they are distributed or working in a single studio. Allowing creative teams to have a clear run at producing their best work in a way thats independent of their location or time zone, is to enable production teams to give the best they have to offer as a distributed workforce. Through ubiquitous access, the cloud offers a wide-reaching tool to streamline the entire process of content storage, production, and distribution by empowering production teams to deliver content from anywhere.

Another unique advantage of cloud storage is its capacity for nearly seamless disaster recovery (DR). The results of outages and hardware or system failure can be extremely costly, both in the short term and long term. In an outage, there is an immediate financial cost to be reckoned with, but there is also reputational harm that can be even more damaging in the long run. Audience loyalty can be considered a long-term investment for media companies one that can disappear in a matter of minutes if something time-sensitive, such as a live sports event or a severe weather warning, fails to reach its intended audience. Extreme weather events, power outages, and ransomware attacks are all looming possibilities for broadcasters and since they arent going away, the only pragmatic approach for media companies who wish to keep and grow their audiences is to treat these eventualities as certain outcomes to prepare for.

With the increased prevalence of multi-distribution endpoints in todays media landscape, seamless delivery of assets to the required endpoints has become a daunting task. The clouds flexibility and scalability make it an ideal solution to accommodate the demands of multi-distribution endpoints. Cloud-based distribution provides a flexible framework that accommodates the complexity of content delivery in todays media landscape allowing M&E companies to adapt to changing demands and technical challenges with a minimum of friction.

The full extent of the clouds archiving capabilities may not be obvious. Still, an overview of its varied use cases, with everything from disaster recovery to creative collaboration, suggests that the cloud can be usefully thought of as a toolset as opposed to a single tool for one purpose. On the other hand, the cloud comes as a single all-in-one solution making it one of the most versatile and budget-friendly investments for broadcasters looking to drive their business operations to the next level.

AboutSam Peterson, BitcentralAs Chief Operating Officer Sam Peterson has direct control of the organizations operations in accordance with Bitcentrals strategic objectives and business plans. Sam is a career broadcast professional with more than 32 years of experience including over 20 years in various leadership roles at large, industry-leading corporations. He has held positions spanning product management, marketing, direct sales, and sales engineering. His broad experience spans all aspects of media production and distribution including automation, master control and playout, news gathering and production, post-production, signal management, and live production. He has successfully overseen the integration of several products from international acquisitions into the North American market. On the OTT side, Sam has extensive networking expertise in contribution and distribution networks for live streaming and video on demand.

Read more here:
More than meets the eye: The full value of cloud archiving - NewscastStudio

Read More..

Five reasons to invest in cloud solutions for your law firm – Lexology

According to 2022Wolters Kluwers Future Ready Lawyerstudy based on interviews of 751 lawyers in Europe and the USA, 82% of law firms expect greater use of technology in the coming years to enhance productivity.

Many law firms have come to realise that the effort to remain competitive means that they need to quickly adapt to changes in the market by fully using technology. Therefore, afuture-proof IT infrastructureplays a crucial and empowering role in this direction, enabling you to meet rising expectations from your clients and employees. Modern IT solutions can help successfully master the challenges of an evolving and more complex legal landscape, by optimising workflows and fostering productivity.

Today, more and more law firms are investing in modern IT and specifically, incloud-based solutions, aware of the several advantages the latter offer over traditional, old-fashioned IT infrastructures.

Here are the5 top reasonsfor whichyou should consider investing in modern cloud solutionsfor your law firm:

Solid IT equipment helps your law firm stay up to date and competitive. After all, those who use outdated technologies run the risk of losing touch and falling behind the competition. As law firms can work faster and more efficiently by applying modern IT systems and using cloud, they can significantly boost their competitiveness.

Optimising work processes in your law firm translates into faster and more effective communication with your clients. For instance, you can offer your clients quick and easy access to information and effectuate clients transactions with far greater speed and security.

Another important aspect that touches upon client needs involves data security. As law firms have access to sensitive information, they must take steps to ensure that it is protected against unauthorised access, protecting the law firms reputation and clients trust towards it. Modern IT solutions offer many possibilities in this regard, including encrypted data transmission and cloud storage that meet the strictest standards in security, thus avoiding reputational harm, loss of clients, payments of ransom and penalty fees that a cyberattack can bring about.

With many processes increasingly being handled online and thus huge amount of data that need to be transmitted quickly and securely, ensure to be compliant with strict legal regulation can be challenging. Modern IT systems can help you meet these requirements and simplify your work. One example is the General Data Protection Regulation (GDPR), which institutes high requirements in terms of data protection. E-mail encryption or the use of secure cloud solutions can also help protect personal data in a secure and reliable manner.

Introducing ultramodern technologies can simplify and automate workflows, saving valuable resources for a law firm, such as time and costs. For example, with the availability of cloud storage, the laborious and time-consuming task of manually searching for documents in the law firm's archives is no longer necessary. Using cloud solutions always gives lawyers access to vital information, from any location; this shortens the time they need to respond to enquiries, thanks also to improved collaboration and productivity (e.g., project management tools, shared calendar, etc.)

Finally,cloud solutions are also often more cost-effectivethan traditional IT systems, as they do not need high investment costs and instead are paid in the form of a monthly subscription.

Would you like to learn all the other benefits of adopting modern IT and cloud solutions? Check our dedicated whitepaper and transform your law firm into a modern cloud-based one!

Continued here:
Five reasons to invest in cloud solutions for your law firm - Lexology

Read More..

Amazon’s Adobe Creative Cloud deal is out of this world – TrustedReviews

If youre looking to save big on Adobes Creative Cloud suite, you wont want to miss this great Amazon offer. The retailer has slashed a huge 32% off the price of a 1-year subscription to Adobe Creative Cloud in its Prime Big Deal Days October sale.

This means that Prime members can bag a years access to 20+ desktop apps for Mac and PC for just 371.41. Thats a 177.37 saving over the course of the year compared to Creative Clouds usual 548.78 price tag.

All you need to take advantage of this offer is an Amazon Prime account, which new members can try for a month free of charge with Amazons 30-day free trial.

Save 32% on 12 months of Adobe Creative Cloud

Gain access to 20+ Adobe desktop and mobile apps (including new generative AI features) for 371.41 with this huge Creative Cloud bundle. Shop now to save 177.37 on Photoshop, Premiere Pro and more.

Creative Cloud is an Adobe software bundle that encompasses more than 20 desktop and mobile apps designed to aid creativity. Popular programs include Photoshop, Illustrator, Premiere Pro, InDesign, Acrobat Pro and Adobe Express.

On top of this, subscribers gain access to more than 1000 free fonts, 100GB of cloud storage for projects and files and a range of tutorials, inspiration and community support.

The software is installed on your computer meaning you have the ability to work offline, while automatic updates ensure your apps are consistently up to date.

Creative Cloud also includes new AI features powered by Adobe Firefly such as Generative Fill in Photoshop. This makes it possible to remove, replace and generate images with a simple text prompt for faster and easier image manipulation.

If youve been thinking about investing in Adobes creative tools, you dont want to miss out on this offer. Head to Amazon now to save 32% on a years subscription to Adobe Creative Cloud and get 20+ apps for as little as 371.41 down from 548.78.

Original post:
Amazon's Adobe Creative Cloud deal is out of this world - TrustedReviews

Read More..

Freakonomics author: ‘Objections to data science in K-12 education … – Yahoo Finance

The three-year battle over Californias new math framework has produced calamity and confusion on all fronts. As fighting raged across op-ed pages and X (formerly known asTwitter), the fog of war obscured the inescapable truth: The data revolution is here and our kids are not prepared for it.

From ChatGPT to personal finance, nearly every decision we make in our daily lives is now dominated by data. Eight out of ten of the fastest-growing careers this year involve data science. A decade from now, it will be difficult to find any job that is not data-driven.

We need to equip our students for this new reality by teaching them basic data literacy in K-12. We can all see this, but somehow the politics of the moment have turned this idea into a raging debate.

The new critics of data science instruction seem to have three common objections. Their first claim is that data science programs are somehow watering down math. That is indeed possible, especially if districts treat data-related classes as a form of remediation, but this should not be the case. Data science is a very challenging subject, combining traditional math, statistics, computer programming, and complex datasets. In many ways, it demands more of students, requiring critical thinking, creativity, and a nuanced understanding of the context within which data have been generated.

A second objection is that learning data science in high school is somehow illegitimate because students wont yet have the mathematics skills required of professional data scientists. This is an odd argument. Can high school students never learn anything about physics because they dont understand differential calculus? Can they not find beauty in a Shakespearean sonnet if they dont know the rules of iambic pentameter?

The third claim is that data science coursework will crowd out calculus or some of the other math required for college STEM degrees. This is an important concern, but it assumes that every part of todays curriculum is absolutely critical to that path. Do we really think that is true? Having spent many nights at the kitchen table helping my kids with their homework, I suspect its not. And we (parents) shouldnt ignore the more than 130 college disciplines that now require data and statistics basics as the world changesincluding math and engineering.

Story continues

We adults can stand around and dither, but young people are not waiting for us to figure this out. In college, students are rushing toward data science courses with astonishing speed. The number of data science undergraduate degree programs has exploded nationallyandin every state. At the University of Wisconsin, Madison, it has quickly becomethe fastest-growing major. Not to be outdone, UC Berkeley recently launchedan entire collegededicated to the subject. Our own institution, the University of Chicago, has hired 25 faculty in data science to keep up with student demand.

Sixteen other states have already officially launched or recommended data science in K-12. Some are creating full-year courses, while others are completely redesigning their math pathways. Leading STEM high schools throughout the country are teaching their students the UC Berkeley Data8 program, one of the best collegiate data science courses in the country. Just recently, a group of AP Statistics teachersorganizeda national data science challenge that attracted more than 5,000 students.

Without leadership from policymakers and educators, this revolution will still happen, but the benefits will go disproportionately to the students who are already advantaged. Wealthy parents and tech employees will teach their kids these skills through summer and after-school programs. Is this what we want? Or do we want to ensure that every child gets at least a basic level of data literacy?

If this all rings true to you, if you believe that a modern K-12 education requires at least some data science instruction, then you can help move us toward action. Ask your local school to incorporate data across school subjects throughout K-12. Ask your teachers to bring modern data tools into the way they teach. And ask your school leaders to offer data-focused math coursesand support their educators with the right resources to do so.

Lets put down our weapons in this math war and start fighting again for our kids futures.

StevenLevitt is an economist, the founder of The Center for Radical Innovation for Social Change (RISC) at UChicago, and the author of Freakonomics.

The opinions expressed in Fortune.com commentary pieces are solely the views of their authors and do not necessarily reflect the opinions and beliefs ofFortune.

This story was originally featured on Fortune.com

Follow this link:

Freakonomics author: 'Objections to data science in K-12 education ... - Yahoo Finance

Read More..

Data Science Program and Department of Music Launch Industry … – University of Arkansas Newswire

Photo Submitted

The Department of Music, in partnership with theData Science Program, recently launched an innovative, first-of-its-kind music industry concentration.

This unique concentration is the only one of its kind nationally, and it incorporates music into the field of data science to create a multidisciplinary curriculum that allows students to explore both their artistry and technical literacy.

Jake Hertzog, assistant professor of guitar and jazz area coordinator, wantedDepartment of Musicstudents to have the opportunity to explore the business side of the music industry. He also saw the need to open creative programs up to non-performers.

"This new concentration broadens the scope of music education," Hertzog said. "It allows students interested in music to be industry leaders, growing alongside an increasingly technical field and supporting our mission of fostering 21st century leaders in the musical world."

Students of this program will graduate with a B.S. in data science with a concentration inMusic Industry Data Analytics (MIDA). While all data science students take the same core courses, each student is also asked to officially declare a concentration, like MIDA, which then functions like a built-in minor.

Several students have already shown interest in the MIDA concentration, including Breck Husong, a sophomore data science student from Cave Springs.

"The MIDA concentration is actually what drew me to the Data Science Program," Husong said. "As a lifelong musician who never had plans on making a career out of performance, the arrival of the MIDA concentration was a pleasant surprise.I really enjoy programming, so the chance of being able to still program while being connected to the music industry was very appealing to me."

Karl Schubert, associate director of the Data Science Program, was the one to take MIDA from concept to reality and said the vision for the concentration is perfectly suited for students with interests similar to Husong's.

"We are excited about the world of opportunities that this collaboration will bring," Schubert said. "It is the goal of our program to prepare students for all types of careers, including those within the high-tech field of music."

Additionally, the 21-credit hour MIDA concentration allows students to take a variety of courses related to music production, as well as introductory business and data-mining courses that help students to succeed in the modern job market.

The Data Science Program offers a B.S. in data science with a multitude of concentrations through the combined efforts of the UofA'sCollege of Engineering, theFulbright College of Arts and Sciences and theSam M. Walton College of Business. Instructors from across campus likewise collaborate to share their individual expertise through the programs many course offerings.

To learn more about the Data Science Program, visitdatascience.uark.edu.

Read the original here:

Data Science Program and Department of Music Launch Industry ... - University of Arkansas Newswire

Read More..

LLNLs Data Science Summer Institute hosts student interns from … – Lawrence Livermore National Laboratory (.gov)

Lawrence Livermore National Laboratorys Data Science Summer Institute (DSSI) hosted summer student interns from Japan on-site for the first time, where the students worked with Lab mentors on real-world projects in artificial intelligence (AI)-assisted bio-surveillance and automated 3D printing.

From June to September, the three students Raiki Yoshimura, Shinnosuke Sawano and Taisei Saida lived in rental apartments near the Lab and worked at the Lab on different data science projects using electronic health records and neural networks trained on experimental data.

Sponsored by Japans Agency for Medical Research and Development (AMED) and the Japan Science and Technology Agency (JST), the relationship between the Data Science Institute (DSI) and the Ministry of Japan stems from a series of agreements that arose in the wake of the Fukushima nuclear accident, to conduct academic exchanges and expand scientific collaboration. The DSSI opportunity began in 2019 but had to go all-virtual due to the COVID-19 pandemic. This summer was the first time that students from Japan were brought on-site.

DSI Director Brian Giera said the relationship is part of a continuing trend of Livermore positioning itself to partner with Japan via scientific connections and one the Lab hopes to expand in the coming years.

The Data Science Institute has found itself an example of the U.S.s posture in using science and technology to partner with geopolitically relevant allies, Giera said. Livermore is establishing itself as a global leader in data science, and its very clear that Japan has oriented in the direction of producing candidates that are highly attractive to us. The collaboration is showing that we are helping the students steer their curricular or scientific focus and sharing the messaging to academia [in Japan] that indeed, data science is relevant. Livermore can help be a partner in realizing that vision, and students are the boots on the ground of having that occur.

Yoshimura, Sawano and Saida had never been to the U.S. before arriving in Livermore for their internships. The DSSI program helped the students find housing and matched them with mentors based on the students technical knowledge. Although the students had almost no knowledge of LLNL prior to their internships (outside of photos on the Internet) and were challenged with cultural and language barriers, they were able to acclimate to working in a national laboratory environment and had successful experiences.

While the students said it took some time to get used to the Labs security procedures and American food, they said they found working with their mentors rewarding and eye-opening.

Sawano, a Ph.D. student at the University of Tokyo with a clinical background in cardiology and treating patients with cardiovascular disease, said he thought his internship was a perfect opportunity to combine his knowledge of data science and medicine. His work is supported in Japan by AMED, and though he didnt know what national labs did prior to his internship, he considers LLNL a fascinating option for a career after he finishes his Ph.D.

Working with his lead mentor, LLNL computer scientist Priyadip Ray, and Lab researchers Andre Goncalves and Jose Cadena-Pico, Sawano applied machine learning to electronic health records obtained from Kaiser Permanente, for a bio-surveillance project funded by the Department of Homeland Security. The Lab is developing AI tools to perform faster diagnostics to allow scientists to detect biological threats earlier, thus providing more time to develop possible countermeasures, according to Ray.

As these tools advance, and if we are able to look at the clinical record of everybody in this country in real-time, then we can detect these kinds of anomalies much faster, Ray said. That would give us more time for developing countermeasures.

Sawano said he hopes the research can someday make it into clinical practice.

Collaborating with the professionals on the team is a really good experience for me because Im usually analyzing data by myself and I check my data by myself, but in this team, Priyadip, Jose and Andre can check my coding and my results. As a result, our output is bigger than I do myself, Sawano said. We were really fortunate to get great mentors; they gave me a great deal of advice and support. It will take time to deliver our results to society, but Im excited about the potential output of our research.

Ray, who also worked with Yoshimura on a project using neural networks to evaluate and predict the impact of gene interactions on the viral load of the HIV virus, said the students did a remarkable job in contributing to advancing the research, which could have concrete impacts to public health in the future.

Yoshimura, who attends Nagoya University and studies biology, said he recently began using machine learning to predict clinical outcomes and was drawn to the DSSI internship to expand his skill set and apply his knowledge to large-scale datasets.

This internship experience has been great because they have very big data sets here that I can apply deep learning to, Yoshimura said. In Japan, we have to pay for data and the data is very little so I cant apply a graph or something like that.

Saida, who studies civil engineering in Japan and wants to be a university professor, worked with his mentor LLNL postdoctoral researcher Aldair Gongora on a project on self-driving labs for additive manufacturing, where machine learning approaches are used to help decide which experiments to do next to speed up manufacturing processes with the goal of eventually deploying these approaches on fully automated robotic systems.

There has been a big push toward autonomous experimentation or self-driving labs, and I think that our work really puts us in a position to continue contributing to those fields in a way that really adds value to the modeling and decision-making components of these systems, said Gongora, who works in the Analytics for Advanced Manufacturing group in the Materials Engineering Division. With all the tools that we now have in data science, its blurring the lines between data scientists, chemists, physicists and engineers.

Gongora said he found Saidas background in programming, machine learning and data science a perfect fit for the project and marveled at how his mentee was able to pick up algorithms and implement them at an extremely fast pace.

Saida said he learned how to use and program robots and integrate the hardware with the software on 3D printing machines, and that he found his internship valuable in expanding his knowledge of Bayesian optimization and mechanical engineering.

Its the first time for me researching outside of Japan, so its been a great experience to research with people in the U.S., Saida said. I had a really positive experience working on these projects, especially 3D printing.

During their summer in Livermore, the students explored the Bay Area, enjoying local pizza and ramen restaurants, seeing tourist attractions and even taking a trip to Yosemite National Park. They also attended regular DSSI community events, meet-ups and ice cream socials, where they got to know their fellow intern cohort.

The mentors said they found the experience just as valuable as the students did. Gongora, who came to the Lab as a foreign national himself, said he resonated with the cultural challenges the students faced and said the opportunity epitomized the strength of diversity in science.

The benefit for me has really been being able work with someone from another country and learning more about Japan; learning about their lives, how their academic journey differs from education here in the U.S., and really finding the commonalities and differences, Gongora said. I'm taking away a lot of new perspective leveraging the expertise that [Saida] was able to bring to the project, both in terms of how very skilled he was at the data science concepts and the brainstorming of new ideas from the civil engineering perspective. Time and time again, the Lab teaches me that it's really through the diversity in thought that these interdisciplinary and multidisciplinary ideas emerge, and I think were stronger for it.

Fellow mentor Ray added that the experience was an incredible opportunity for the Lab to get some of the best students from Japan and that he looks forward to mentoring more students from that country.

At the Lab, we are trying to solve very impactful and challenging problems, and we want the best teams to work on these problems, Ray said. There are communication barriers and cultural barriers, because many of them are coming for the first time, but once they're on site, I see them working very hard to overcome all the challenges and really contribute to the mission and push things forward.

Continued here:

LLNLs Data Science Summer Institute hosts student interns from ... - Lawrence Livermore National Laboratory (.gov)

Read More..

I Write About Science and Technology -Part 1: My Main Content on … – Medium

Photo by Scott Graham on Unsplash

Recently, Ive started having job requests on writing about science and technology, one client focused on biology, bioinformatics, and AI to assist these fields; and the other about AI language models and their applications. I thought it would be good to summarize what kinds of contents Ive so far written here at Medium, where the choice of topic and style is all mine -and then I can adapt to your needs. Heres a summary of my topics about data science, AI, numerical analysis and programming.

As a passionate technologist, scientist, writer, and technology integrator as I call myself, my articles on Medium span the fascinating intersection of data science, artificial intelligence, numerical analysis, and programming -oftentimes mixed with chemistry and biology, though not always, and sometimes touching on various technologies such as virtual reality or blockchains, just two mention two.

I try to write each piece as a journey into the heart of these rapidly evolving fields, offering insights into their latest advancements and applications. All making part of the LucianoSphere.

One of my key interests lies in the realm of AI and its transformative impact on various disciplines. Ive written extensively about DeepMinds AlphaFold 2 and its revolutionary role in protein research, as well as the potential of AI in discovering new antibiotics. The philosophical aspects of AI, especially of large language models, such as the Turing Test and the Chinese Room Argument, also feature prominently in my work, offering a deeper exploration of this technologys implications. Ive written about the disruptive potential of huge protein language models in, say, the classroom, and a lot also in biology.

Ive delved into numerical integration, fitting, Monte Carlo simulations, and much more, and their applications in natural sciences, engineering, and economics. My articles often discuss the practical use of these mathematical tools in simplifying equation modeling tasks, such as the Michaelis-Menten equation for enzymatic catalysis.

View post:

I Write About Science and Technology -Part 1: My Main Content on ... - Medium

Read More..

Why SQL is THE Language to Learn for Data Science – KDnuggets

Python!No, R.Fools, its obviously Rust.

Many data science learners and experts alike are keen to pin down the very best language for data science. In my opinion, most people are wrong. Amidst the hunt for the newest, the sexiest, the most container-able data science language, people are looking for the wrong thing.

Its easy to overlook. Its easy to even discount it as a language. But the humble Structured Query Language, or SQL, is my pick for the language to learn for data science. All those other languages certainly have their place, but SQL is the one non-negotiable language that I consider a base requirement for anyone working in data science. Heres why.

Look, databases come hand in hand with data science. Its in the name. If youre working with data science, youre working with databases. And if youre working with databases, youre probably working with SQL.

Why? Because SQL is the universal database query language. There is no other. Imagine someone told you that if you just learned a specific language, youd be able to speak to and understand every single person on Earth. How valuable would that be? SQL is that language in data science, the language that everyone uses to manage and access databases.

Every data scientist needs to access and retrieve data, to explore data and build hypotheses, to filter, aggregate, and sort data. And hence, every data scientist will need SQL. As long as you know how to write a SQL query, youll go far.

Someone, reading this article right now, is piping up about the NoSQL movement. Indeed, certain data is now more commonly stored in non-relational databases, such as by key-value pairs or graph data. Its true that there are benefits to storing data like that you gain more scalability and flexibility. But theres no standard NoSQL query language. You might learn one for one job, and then need to learn an entirely new one for a new job.

Plus, you will very rarely find a business that works entirely with NoSQL databases, while many companies dont need non-relational databases.

Theres that famous (and debunked) stat about how data scientists spend 80% of their time cleaning. While its not true, I think if you ask any data scientist what they spend time on, data cleaning will rank in the top five tasks. Thats why this section is the longest.

You can clean and process data with other languages, but SQL in particular offers unique advantages for certain aspects of data cleaning and processing.

SQL's expressive query language allows data scientists to efficiently filter, sort, and aggregate data using concise statements. This level of flexibility is especially useful when dealing with large datasets where manual data manipulation would be time-consuming and error-prone. Compare that to a language like Python, where achieving similar data manipulation tasks might require writing more lines of code and dealing with loops, conditions, and external libraries. While Python is renowned for its versatility and rich ecosystem of data science libraries, SQL's focused syntax can expedite routine data cleaning operations, enabling data scientists to swiftly prepare data for analysis.

Plus, any data scientist will complain about the bane of their existence: missing values. SQL's functions and capabilities for handling missing valuessuch as using COALESCE, CASE, and NULL handlingprovide straightforward approaches to address gaps in data without the need for complex programming logic.

The other bane of a data scientists existence is duplicates. Happily, SQL offers efficient methods to identify and eliminate duplicate records from datasets, like the `DISTINCT` keyword and the `GROUP BY` clause.

Youve probably heard of ETL pipelines. Well, SQL can be used to create data transformation pipelines, which take raw or semi-processed data and convert it into a format suitable for analysis. This is particularly beneficial for automating and standardizing that repetitive data-cleaning processes we all know and hate.

SQL's ability to join tables from different databases or files streamlines the process of merging data for analysis is essential for projects involving data integration or aggregating data from diverse origins. Which, for a data scientist, comprises a majority of projects.

Finally, I like to remind people that data science does not happen in a vacuum. SQL queries are self-contained and can be easily shared with colleagues. This fosters collaboration and ensures that others can reproduce data cleaning steps without manual intervention.

Now, you wont get far in data science if you only know SQL. But happily, SQL integrates perfectly well with any other of the top data science languages like R, Python, Julia, or Rust. You get all the benefits of analysis, data viz, and machine learning while still retaining SQLs strength for data manipulation.

This is especially powerful when you think about all that data cleaning and processing I talked about earlier. You can use SQL to preprocess and clean data directly within databases, and then lean on Python, R, Julia, or Rust to perform more advanced data transformations or feature engineering, leveraging the extensive libraries available.

Many organizations rely on SQL or, more accurately, rely on data scientists who know how to use SQL to generate reports, dashboards, and visualizations that inform decision-making. Familiarity with SQL enables data scientists to produce meaningful reports directly from databases. And because SQL is so widespread, these reports are usually compatible and interoperable across almost any system.

Because of how interoperable it is with reporting tools and scripting languages like Python, R, and JavaScript, data scientists can actually automate the reporting processes, seamlessly combining SQL's data extraction and manipulation capabilities with the visualization and reporting features of these languages. The upshot is you get comprehensive and insightful reports that effectively communicate data-driven insights to stakeholders, all inside one place.

Theres a reason youll get asked a bunch of SQL interview questions at any data science interview. Almost every data science job requires at least a basic familiarity with SQL.

Heres an example of what I mean: the job listing says, Expertise in SQL, and R or Python for data analysis and platform development. In other words, SQL is a must. And then either R or Python, but one is as good as another to most employers. But thanks to SQL domination, theres no alternative to SQL. Every data science job will require you to work with SQL.

The really cool thing about it is that it makes SQL the ultimate transferable tool. One job may prefer Python, while a startup might require Rust due to personal preference or legacy infrastructure. But no matter where you go, or what you do, its SQL or bust. Take the time to learn it, and youll always be able to tick off a job requirement.

Ultimately, if you find a job as a data scientist that doesnt require SQL, youre probably not going to be doing a whole lot of data science.

It really comes down to the database. Data science requires the storage, manipulation, retrieval, and management of a lot of data. That data lives somewhere. It can only be accessed with one tool, normally, and that tool is SQL. SQL is the language to learn for data science and will be for as long as we rely on databases to do data science.

Nate Rosidi is a data scientist and in product strategy. He's also an adjunct professor teaching analytics, and is the founder of StrataScratch, a platform helping data scientists prepare for their interviews with real interview questions from top companies. Connect with him on Twitter: StrataScratch or LinkedIn.

Go here to see the original:

Why SQL is THE Language to Learn for Data Science - KDnuggets

Read More..

Rady School Leverages AI to Prepare Next Generation of Business … – University of California San Diego

Why has GenAI been integrated into the MSBA programs at Rady?

Nijs: GenAI is already transforming the ways people learn and work in many ways. Rather than fight the coming changes, we aim to teach our students skills that will complement what these tools already bring to the table.

We will be able to improve students math, programming and machine-learning skills in less time. As a result, our students will be able to develop a deeper understanding of the business problems they will solve using GenAI and machine learning during their capstone projects (in which they help an organization solve a real and pressing business challenge) and after graduation. Our part-time MSBA students, who are working professionals, will be able to have an immediate impact and bring value to their organizations as they learn to integrate GenAI into existing business processes or design new ones where different tasks can be maximally supported by GenAI.

In addition, GenAI will reduce the technical barriers to be successful in the program. For example, translating ideas into Python code with these tools will be much more efficient. Every incoming MSBA student along with program instructors and teaching assistants, will have access to GenAI to help in task completion. They basically have a personal AI tutor and can spend more time on valuable tasks like data exploration, applying advanced analytics and creating a tangible impact on the business problems they are trying to solve. Developing talent that has this experience, along with strong business acumen, will create business data scientists that are perfectly equipped to solve the business problems of the future and find innovative ways to help businesses grow.

Nijs: The Bureau of Labor Statistics projects data science job growth of 36% by 2031, which is much faster than the average for other roles. This is because businesses have access to huge amounts of data about their customers, supply chains, markets, devices, services, etc. So much so that developing creative ways to access, analyze and find patterns is particularly challenging but also necessary to stay competitive. Not only do business data scientists have to find meaningful insights, but they have to do it fast and at scale. GenAI can help business data scientists achieve these goals.

Business data scientists are integral to helping organizations make informed decisions, improve their business processes, design and develop new products, effectively market their products and more.

Nijs: We see GenAI as both a skill multiplier and a skill extender. That means students offered GenAI-enhanced learning will be stronger in terms of both their depth and breadth of understanding and abilities. By taking this big step forward we will have to work on new and better ways to continue challenging our students. GenAI-assisted learning will allow us to increase class expectations with respect to both the quantity and quality of student work. We are also creating customized GenAI policies for each MSBA class to ensure students are able to leverage these tools appropriately to maximize learning.

Students will also learn to use GenAI responsibly. Checking the work and using GenAI as a partner, but making sure that it's not hallucinating and that the work they submit is valid, credible and returns appropriate results. No matter if GenAI was or wasnt used for a particular task, we require that students can understand and explain every detail of their work. Students need to learn how to partner effectively with GenAI which takes practice.

Nemteanu: These technologies are still so new that the first major experimental studies are just starting to come out. They clearly show the massive impact these tools can have on individual and company performance.

For example, a paper published in Science demonstrated that ChatGPT was able to reduce the time required for a business writing task by 40% while at the same time increasing the quality of the work. After using AI, participants were also significantly happier. Why? Likely because it helped them complete tedious work at a much faster pace.

Our full-time students are going to have the real-world experience of using these technologies, building analytic solutions better and faster, something that will resonate with hiring managers. Students in our part-time MSBA program, who are working professionals, are going to be able to take what they learn and immediately apply it in their organization and have a lasting impact on near-term and long-term strategies and business goals.

Nemteanu: We believe that GenAI-assisted business analytics can add an extremely valuable layer of skills for many UC San Diego graduates. Studying data science for business (aka Business Analytics) is important because business is a science--whether its finance, marketing, supply chain--they require data and analysis to ensure every decision is data driven and credible.

If you have a technical undergraduate degree, we can bolster your job market outcomes (i.e., job title and salary) by helping you translate and enhance your skillset and set you on a GenAI-assisted path that solves business problems using data science tools. If you have strong domain expertise in a substantive area, e.g., biology, we can augment your abilities with a high-powered data and GenAI literate skillset. In short, a masters degree in Business Analytics can supplement any background and contribute to a very exciting future with amplified career opportunities.

Original post:

Rady School Leverages AI to Prepare Next Generation of Business ... - University of California San Diego

Read More..