Page 368«..1020..367368369370..380390..»

Matillion Announces Release of Data Productivity Cloud for Databricks – AiThority

No-code data pipeline platform unlocksDelta Lakefor BI and machine learning

Matillions Data Productivity Cloud is now available for Databricks, enabling users to access the power of Delta Lake within their data engineering.

The Data Productivity Cloud with Databricks brings no-code data ingestion and transformation capabilities that are purpose-built for Databricks, enabling users to quickly build data pipelines at scale, to be used in AI and analytics projects.

Ciaran Dynes, Chief Product Officer at Matillion said:In the months since launching Data Productivity Cloud, weve continued to integrate more cloud platforms and data sources to bring no-code data engineering tools to every member of the data team.

Recommended AI News: Talkdesk Autopilot: AI Customer Service for Banks & Retailers

We know that SQL is a massive workload for Databricks, so as well as unlocking the value of Delta Lake and lakehouses, were excited to bring no-code tooling to the Data Productivity Cloud to help Databricks users be even more productive running SQL pipelines on the Databricks platform.

Data Productivity Cloud enables users to easily build data pipelines with any data source for business intelligence and data science projects.

Matillion has long partnered with the data and AI company Databricks, and Databricks Ventures is an investor in the tech unicorn.

Recommended AI News: Epiq AI Expands Legal Solutions as Part of the Epiq Service Cloud

Matillion delivers a comprehensive data integration solution that can fit seamlessly within the existing tech stack, helping data teams to centralise and transform data from any source into business-driving insights at speed.

Data Productivity Cloud with Databricks offers consolidated billing with Matillions transparent credit-based pricing model, with setup in minutes. Thousands of enterprises including Cisco, Docusign, Slack and TUI trust Matillion to move, transform and orchestrate data for a wide range of use cars from insights and operational analytics, to data science, machine learning and AI.

Recommended AI News: OurCrowd AI Fund to Collaborate with NVIDIA Inception

[To share your insights with us as part of editorial or sponsored content, please write to sghosh@martechseries.com]

Originally posted here:
Matillion Announces Release of Data Productivity Cloud for Databricks - AiThority

Read More..

The Top 3 Machine Learning Stocks to Buy in March 2024 – InvestorPlace

Source: NicoElNino / Shutterstock.com

You may be hearing the word AI bubble a lot these days, especially regarding the stock market. Since OpenAI released its artificial-intelligence (AI) chatbot ChatGPT in Nov. 2022, it feels like every company in the world has been getting into the AI business.

Machine learning is a type of AI that allows computers to learn and reproduce how humans learn and use that to replicate their behaviors. As you might imagine, machine learning has the potential to decrease the cost and time of human tasks and eliminate redundant work.

Companies are set to save billions of dollars by integrating machine learning tools and software in their businesses. As investors, not only is it important to look at which companies are successfully using machine learning, but also the companies that are providing these tools to be used. This article will discuss three of the top machine-learning stocks to buy while the AI industry remains red-hot.

Source: Poetra.RH / Shutterstock.com

NVIDIA (NASDAQ:NVDA) is the global leader when it comes to producing GPUs that can power machine-learning computers. The stock has been on a tear over the past year, returning north of 240% to shareholders, while surging up the list of the worlds most valuable companies. Despite such unprecedented growth, Yahoo Finance analysts still remain optimistic for with with a one-year target between an average of $852.10 to a high of $1,400.0.

When it comes to machine-learning GPUs, NVIDIA is second to none in the semiconductor industry. NVIDIA has more demand for its chips than it has supply even at elevated prices, an its customers include some of the most powerful companies in the world.

You might think that a stock that has risen by more than 240% in one year is overinflated. The fact is, that NVIDIAs revenue has grown so fast that its growth has kept pace with its stock valuation. Looking comparatively, NVIDIAs forward P/E ratio of 34.25x is still lower than the likes of Amazon and Tesla. As long as AI and machine learning are being adopted, NVIDIAs stock should continue to reap rewards for investors.

Source: Roschetzky Photography / Shutterstock.com

Tesla (NASDAQ:TSLA) is a company that needs no introduction. It is the largest manufacturer of electric vehicles in the world and single-handedly revolutionized the auto industry. While its stock has lagged behind its other Magnificient 7 counterparts in 2024 due to high-interest rate environments, its consensus one-year price target still aims for a high of $345.00.

So, how does an electric vehicle company operate in the machine-learning industry? Tesla, led by CEO Elon Musk, has long been trying to master self-driving technology. Teslas FSD or full self-driving software has had some roadblocks from regulatory agencies like the NHTSA in America, but Musk remains confident that it will be available to all Tesla users in the future.

Teslas stock still trades at a premium, especially since the company has reported declining operating margins and fairly stagnant revenue growth. The forward P/E ratio of the stock shows that TSLA is trading at about 65x forward earnings, which is nearly double that of NVIDIA. As mentioned, Teslas stock could continue to struggle until interest rates begin to decline. Savvy long-term investors might be taking this period of consolidation as a time to load up on the high-growth stock.

Source: Mamun sheikh K / Shutterstock.com

Palantir (NYSE:PLTR) is a data analytics and software company that has a very polarizing following on social media. At one time, Palantir was looked at as a meme stock, but the company has since proven to be profitable and has exhibited impressive growth.

While the operations of Palantir have always been shrouded in mystery, the company has made clear progress in growing its customer base over the past few years. One of the ways it has done this is by introducing its AIP or Artificial Intelligence Platform. AIP uses machine learning to help large-scale enterprises unluck insights from large sets of data. From this analysis, companies can identify inefficiencies and operate at a higher level.

We did mention Palantirs stock is trading at the high end of analyst estimates, right? Well, although it is a much smaller company, Palantirs valuation currently dwarfs that of both NVIDIA and Tesla. At its current price, Palantirs stock trades at about 25x sales and 79x future earnings. With the potential to be considered for S&P 500 inclusion later this year, and management guiding a FY2024 revenue of around $2.6 billion, Palantir is a worthy company to look into capitalizing off machine learning.

On the date of publication, Ian Hartana and Vayun Chugh did not hold (either directly or indirectly) any positions in the securities mentioned in this article. The opinions expressed in this article are those of the writer, subject to the InvestorPlace.comPublishing Guidelines.

Chandler Capital is the work of Ian Hartana and Vayun Chugh. Ian Hartana and Vayun Chugh are both self-taught investors whose work has been featured in Seeking Alpha. Their research primarily revolves around GARP stocks with a long-term investment perspective encompassing diverse sectors such as technology, energy, and healthcare.

Continued here:
The Top 3 Machine Learning Stocks to Buy in March 2024 - InvestorPlace

Read More..

New AI logistics tool enhances warehouse efficiency – Imaging and Machine Vision Europe

UK-based robotics and data intelligence company Dexory has introduced an AI-powered logistics engine to help warehouses maximise operational efficiency, optimise inventory management, and enhance the overall warehouse agility and responsiveness.

DexoryView enables warehouse operatives to process millions of data sets captured daily across warehouses via autonomous robots, providing real-time access to insights and predictive analytics that enable businesses to make smarter, faster decisions.

The combination of computer vision, machine learning, natural language processing and large language models allows warehouse operators to maximise efficiency across critical drivers such as space utilisation, inventory, working time and machinery utilisation.

The new AI functionality unlocks the next level of intelligence in DexoryView. The platform combines the use of autonomous robots to scan warehouses of one million square feet (approx 93,000m2) and more than 100,000 pallets in a day, providing accurate, instant, real time information on goods and assets across the sites it operates in.

Credit for main image: dreamnikon/Shutterstock

See more here:
New AI logistics tool enhances warehouse efficiency - Imaging and Machine Vision Europe

Read More..

The ML Product Manager: Building AI-powered Solution – hackernoon.com

The ML Product Manager: Building AI-powered Solution by@ therealsjr

The intersection of machine learning (ML) and product management is a rapidly evolving field. As businesses realize the power of AI-powered solutions across industries like advertising, finance, and healthcare, product managers who understand ML are becoming indispensable. How can product managers build successful careers in this exciting space, grab the unique opportunities ML provides for product innovation, and overcome the challenges of responsibly developing and deploying AI products? I spoke to Gleb Sinev, an ML product manager with experience at companies like Surfingbird, Handl, and Mantika.ai - a startup that is developing software for the early detection of lung cancer through computer vision technology - who offered valuable insights into the ML product manager's journey. Here's what we can learn from his extensive career.

Editor-at-Large at Dataconomy, Editor-at-Large at ArcticStartup, Co-founder at ONOISNO - an impact creator house

Receive Stories from @therealsjr

L O A D I N G . . . comments & more!

Here is the original post:
The ML Product Manager: Building AI-powered Solution - hackernoon.com

Read More..

BlueCat Networks Appoints Scott Fulton as Chief Product and Technology Officer – AiThority

An accomplished executive, he will lead the next phase of product development and growth

BlueCat Networks, a leading provider of secure and automated DNS, DHCP, and IP address management (collectively known as DDI) services and solutions for mission-critical networks, announced that Scott Fulton has joined the company as its new Chief Product and Technology Officer (CPTO).

Fulton brings more than 25 years of enterprise software product management and engineering leadership experience to the role. Recently, he was a founder and CEO of venture-backed cloud observability start-up OpsCruise. Under his leadership, OpsCruise developed patented artificial intelligence and machine learning technology, won Fortune 500 customers, and was recognized by Gartner as a Cool Vendor. The company was acquired in June 2023.

Recommended AI News: Talkdesk Autopilot: AI Customer Service for Banks & Retailers

Prior to that, Fulton was an executive vice president at Infoblox. While in that role, he was responsible for products and corporate development during a transformational period that included the introduction of several security, cloud, and service provider offerings. Earlier in his career, Fulton held general management roles at BladeLogic/BMC Software and Hewlett Packards management software group.

Scotts proven track record in driving innovation and scaling product management aligns well with our plans to greatly accelerate growth, said BlueCat CEO Stephen Devito. Under Scotts leadership, ourindustry-recognized portfoliowill continue to deliver the flexibility and control organizations need to modernize and optimize their networks for change.

Recommended AI News: Epiq AI Expands Legal Solutions as Part of the Epiq Service Cloud

Im thrilled to be joining a world-class organization thats innovating both organically and acquisitively and has a reputation for strong customer and employee engagement, said Fulton. Many of the secular trends in technology are creating new challenges for enterprise networking. Im eager to collaborate with BlueCats talented team, senior leadership, and Audax Private Equity to help drive the company to the next level.

Recommended AI News: Armada and Edarat Group Partner to Introduce Edge Computing and AI to MENAs Industrial Sector

[To share your insights with us as part of editorial or sponsored content, please write to sghosh@martechseries.com]

See more here:
BlueCat Networks Appoints Scott Fulton as Chief Product and Technology Officer - AiThority

Read More..

Identification of diagnostic markers for moyamoya disease by combining bulk RNA-sequencing analysis and machine … – Nature.com

Tinelli, F. et al. Vascular remodeling in moyamoya angiopathy: From peripheral blood mononuclear cells to endothelial cells. Int. J. Mol. Sci. 21(16), 5763 (2020).

Article CAS PubMed PubMed Central Google Scholar

Kuroda, S. & Houkin, K. Moyamoya disease: Current concepts and future perspectives. Lancet Neurol. 7(11), 10561066 (2008).

Article PubMed Google Scholar

Goto, Y. & Yonekawa, Y. Worldwide distribution of moyamoya disease. Neurol. Med. Chir. 32(12), 883886 (1992).

Article CAS Google Scholar

Kuriyama, S. et al. Prevalence and clinicoepidemiological features of moyamoya disease in Japan: Findings from a nationwide epidemiological survey. Stroke 39(1), 4247 (2008).

Article PubMed Google Scholar

Kamada, F. et al. A genome-wide association study identifies RNF213 as the first Moyamoya disease gene. J. Hum. Genet. 56(1), 3440 (2011).

Article CAS PubMed Google Scholar

Bang, O. Y., Fujimura, M. & Kim, S. K. the pathophysiology of moyamoya disease: An update. J. Stroke 18(1), 1220 (2016).

Article PubMed PubMed Central Google Scholar

Kim, E. H. et al. Importance of RNF213 polymorphism on clinical features and long-term outcome in moyamoya disease. J. Neurosurg. 124(5), 12211227 (2016).

Article CAS PubMed Google Scholar

Kang, H. S. et al. Plasma matrix metalloproteinases, cytokines and angiogenic factors in moyamoya disease. J. Neurol. Neurosurg. Psychiatry 81(6), 673678 (2010).

Article PubMed Google Scholar

Jaipersad, A. S. et al. The role of monocytes in angiogenesis and atherosclerosis. J. Am. Coll. Cardiol. 63(1), 111 (2014).

Article CAS PubMed Google Scholar

Morishita, R. et al. Impairment of collateral formation in lipoprotein(a) transgenic mice: Therapeutic angiogenesis induced by human hepatocyte growth factor gene. Circulation 105(12), 14911496 (2002).

Article CAS PubMed Google Scholar

Schning, M. et al. Antiphospholipid antibodies in cerebrovascular ischemia and stroke in childhood. Neuropediatrics 25(1), 814 (1994).

Article PubMed Google Scholar

Suzuki, S. et al. Moyamoya disease complicated by Graves disease and type 2 diabetes mellitus: Report of two cases. Clin. Neurol. Neurosurg. 113(4), 325329 (2011).

Article PubMed Google Scholar

Wanifuchi, H. et al. Autoimmune antibody in moyamoya disease. No Shinkei Geka 14(1), 3135 (1986).

CAS PubMed Google Scholar

Lin, R. et al. Clinical and immunopathological features of Moyamoya disease. PLoS ONE 7(4), e36386 (2012).

Article ADS CAS PubMed PubMed Central Google Scholar

Fujimura, M. et al. Increased serum production of soluble CD163 and CXCL5 in patients with moyamoya disease: Involvement of intrinsic immune reaction in its pathogenesis. Brain Res. 1679, 3944 (2018).

Article CAS PubMed Google Scholar

Kanehisa, M. Toward understanding the origin and evolution of cellular organisms. Protein Sci. 28(11), 19471951 (2019).

Article CAS PubMed PubMed Central Google Scholar

Kanehisa, M. & Goto, S. KEGG: Kyoto encyclopedia of genes and genomes. Nucleic Acids Res. 28(1), 2730 (2000).

Article CAS PubMed PubMed Central Google Scholar

Kanehisa, M. et al. KEGG for taxonomy-based analysis of pathways and genomes. Nucleic Acids Res. 51(D1), D587-d592 (2023).

Article CAS PubMed Google Scholar

Langfelder, P. & Horvath, S. WGCNA: An R package for weighted correlation network analysis. BMC Bioinform. 9, 559 (2008).

Article Google Scholar

Degenhardt, F., Seifert, S. & Szymczak, S. Evaluation of variable selection methods for random forests and omics data sets. Brief Bioinform. 20(2), 492503 (2019).

Article PubMed Google Scholar

Friedman, J., Hastie, T. & Tibshirani, R. Regularization paths for generalized linear models via coordinate descent. J. Stat. Softw. 33(1), 122 (2010).

Article PubMed PubMed Central Google Scholar

Newman, A. M. et al. Robust enumeration of cell subsets from tissue expression profiles. Nat. Methods 12(5), 453 (2015).

Article CAS PubMed PubMed Central Google Scholar

Huang, J. et al. Weighted gene co-expression network analysis and CIBERSORT screening of key genes related to m6A methylation in Hirschsprungs disease. Front. Genet. 14, 1183467 (2023).

Article CAS PubMed PubMed Central Google Scholar

Asselman, C. et al. Moyamoya disease emerging as an immune-related angiopathy. Trends Mol. Med. 28(11), 939950 (2022).

Article PubMed Google Scholar

Sigdel, T. K. et al. Immune response profiling identifies autoantibodies specific to Moyamoya patients. Orphanet. J. Rare Dis. 8, 45 (2013).

Article PubMed PubMed Central Google Scholar

Jin, F. & Duan, C. Identification of immune-infiltrated hub genes as potential biomarkers of Moyamoya disease by bioinformatics analysis. Orphanet. J. Rare Dis. 17(1), 80 (2022).

Article PubMed PubMed Central Google Scholar

Roder, C. et al. Common genetic polymorphisms in moyamoya and atherosclerotic disease in Europeans. Childs Nerv. Syst. 27(2), 245252 (2011).

Article PubMed Google Scholar

Achrol, A. S. et al. Pathophysiology and genetic factors in moyamoya disease. Neurosurg. Focus 26(4), E4 (2009).

Article PubMed Google Scholar

Guo, D. C. et al. Mutations in smooth muscle alpha-actin (ACTA2) cause coronary artery disease, stroke, and Moyamoya disease, along with thoracic aortic disease. Am. J. Hum. Genet. 84(5), 617627 (2009).

Article CAS PubMed PubMed Central Google Scholar

Ikeda, E. Systemic vascular changes in spontaneous occlusion of the circle of Willis. Stroke 22(11), 13581362 (1991).

Article CAS PubMed Google Scholar

Sato-Maeda, M. et al. Transient middle cerebral artery occlusion in mice induces neuronal expression of RNF213, a susceptibility gene for moyamoya disease. Brain Res. 1630, 5055 (2016).

Article CAS PubMed Google Scholar

Quintos, J. B., Guo, M. H. & Dauber, A. Idiopathic short stature due to novel heterozygous mutation of the aggrecan gene. J. Pediatr. Endocrinol. Metab. 28(78), 927932 (2015).

CAS PubMed PubMed Central Google Scholar

Gkourogianni, A. et al. Clinical characterization of patients with autosomal dominant short stature due to aggrecan mutations. J. Clin. Endocrinol. Metab. 102(2), 460469 (2017).

Article PubMed Google Scholar

Lin, L. et al. A high proportion of novel ACAN mutations and their prevalence in a large cohort of chinese short stature children. J. Clin. Endocrinol. Metab. 106(7), e2711e2719 (2021).

Article PubMed PubMed Central Google Scholar

Vafaeie, F. et al. ACAN, MDFI, and CHST1 as candidate genes in gastric cancer: A comprehensive insilco analysis. Asian Pac. J. Cancer Prev. 23(2), 683694 (2022).

Article CAS PubMed PubMed Central Google Scholar

Koh, Y. W. et al. Association between the CpG island methylator phenotype and its prognostic significance in primary pulmonary adenocarcinoma. Tumour Biol. 37(8), 1067510684 (2016).

Article CAS PubMed Google Scholar

Kim, S. M. et al. Endothelial dysfunction induces atherosclerosis: Increased aggrecan expression promotes apoptosis in vascular smooth muscle cells. BMB Rep. 52(2), 145150 (2019).

Article CAS PubMed PubMed Central Google Scholar

Jung, K. H. et al. Circulating endothelial progenitor cells as a pathogenetic marker of moyamoya disease. J. Cereb. Blood Flow Metab. 28(11), 17951803 (2008).

Article CAS PubMed Google Scholar

Kim, J. H. et al. Decreased level and defective function of circulating endothelial progenitor cells in children with moyamoya disease. J. Neurosci. Res. 88(3), 510518 (2010).

Article CAS PubMed Google Scholar

Kashem, M. A. et al. The potential role of FREM1 and its isoform TILRR in HIV-1 acquisition through mediating inflammation. Int. J. Mol. Sci. 22(15), 7825 (2021).

Article CAS PubMed PubMed Central Google Scholar

Li, H. N. et al. Elevated expression of FREM1 in breast cancer indicates favorable prognosis and high-level immune infiltration status. Cancer Med. 9(24), 95549570 (2020).

Article CAS PubMed PubMed Central Google Scholar

Karim, R. et al. Human papillomavirus (HPV) upregulates the cellular deubiquitinase UCHL1 to suppress the keratinocytes innate immune response. PLoS Pathog. 9(5), e1003384 (2013).

Article CAS PubMed PubMed Central Google Scholar

Gu, Y. et al. The deubiquitinating enzyme UCHL1 negatively regulates the immunosuppressive capacity and survival of multipotent mesenchymal stromal cells. Cell Death Dis. 9(5), 459 (2018).

Article PubMed PubMed Central Google Scholar

Jain, M. et al. TOP2A is overexpressed and is a therapeutic target for adrenocortical carcinoma. Endocr. Relat. Cancer 20(3), 361370 (2013).

Article CAS PubMed PubMed Central Google Scholar

Olsen, K. E. et al. Amplification of HER2 and TOP2A and deletion of TOP2A genes in breast cancer investigated by new FISH probes. Acta Oncol. 43(1), 3542 (2004).

Article CAS PubMed Google Scholar

Moretti, E. et al. TOP2A protein by quantitative immunofluorescence as a predictor of response to epirubicin in the neoadjuvant treatment of breast cancer. Future Oncol. 9(10), 14771487 (2013).

Article CAS PubMed Google Scholar

Masuda, J., Ogata, J. & Yutani, C. Smooth muscle cell proliferation and localization of macrophages and T cells in the occlusive intracranial major arteries in moyamoya disease. Stroke 24(12), 19601967 (1993).

Article CAS PubMed Google Scholar

Weng, L. et al. Association of increased Treg and Th17 with pathogenesis of moyamoya disease. Sci. Rep. 7(1), 3071 (2017).

Article ADS PubMed PubMed Central Google Scholar

Yamamoto, M. et al. Increase in elastin gene expression and protein synthesis in arterial smooth muscle cells derived from patients with Moyamoya disease. Stroke 28(9), 17331738 (1997).

Article CAS PubMed Google Scholar

Roder, C. et al. Polymorphisms in TGFB1 and PDGFRB are associated with Moyamoya disease in European patients. Acta Neurochir. (Wien) 152(12), 21532160 (2010).

Article PubMed Google Scholar

Here is the original post:
Identification of diagnostic markers for moyamoya disease by combining bulk RNA-sequencing analysis and machine ... - Nature.com

Read More..

Nextech3D.ai Announces Formation of AI Incubator and AI Acquisition… – MarTech Outlook

TORONTO, ON, Canada - Nextech3D.AI (OTCQX: NEXCF) (CSE: NTAR) (FSE: 1SS), a Generative AI-Powered 3D model supplier for Amazon, P&G, Kohls and other major e-commerce retailers announces the launch of its new AI Incubator and AI Development Division, signaling a significant leap forward in its commitment to advancing artificial intelligence technologies. This new venture, set to be spearheaded by a dynamic team of three esteemed AI scientists alongside the visionary leadership of Evan Gappelberg, CEO of Nextech3D.ai, and Hareesh Achi, former Microsoft executive renowned for his pivotal role in digital transformation and technology strategy. This new division marks a major milestone in Nextech3D.ai's growth trajectory and sets the stage for additional potential IPO spin- outs of new AI technology Companies to its shareholders.

The AI Incubator and Development Division is established with the mission to nurture AI solutions that will enhance the human experience across various sectors, including education, healthcare, ecommerce, blockchain, retail, and more. This initiative is a testament to Nextech3D.ai's dedication to pushing the boundaries of what's possible in AI and 3D technology integration for its shareholders.

Evan Gappelberg, celebrated for his foresight and innovation in the tech arena, will provide strategic direction and vision for the division. His leadership has already positioned Nextech3D.ai and its two spin outs Toggle3D.ai and ARway.ai as a frontrunners in the 3D technology landscape.

Joining him in steering this ambitious project is Hareesh Achi, an ex-Microsoft and META executive whose expertise in AI, digital advertising, and operational efficiency has earned him acclaim in the technology community. Achi's proven track record of leading high-impact teams and his strategic vision for AI applications make him an invaluable asset to the Nextech3D.ai leadership team.

The division's core team of AI scientists, handpicked for their pioneering work and contributions to the field of artificial intelligence, will focus on developing AI-driven technologies and solutions that are not only innovative but also ethically responsible and sustainable.

Katyani Singh, Computer Vision/Machine Learning Scientist who obtained her Masters in Computing Science at the University of Alberta. Her research interests lie in Computer Vision and Deep Learning. Amir Salimnia, Computer Vision/Machine Learning Scientist, who is an MSc. Computer Engineering graduate from Western University with more than three years of experience developing machine learning.

Omid Alemi, Machine Learning Researcher and Software Engineer, who holds an MSc in Computer Science from the University of Northern British Columbia and a BSc in Software Engineering from the University of Arak and has 3+ years experience working on developing and deploying machine learning solutions for computer vision and 3D asset creation.

This core team of AI scientists will conduct research and development efforts geared towards creating AI tools and applications that empower businesses and consumers alike, fostering a future where technology and humanity intersect seamlessly.

New Path to Innovation and Growth

In a groundbreaking move, Nextech3D.ai has announced that once the AI technologies incubated within the AI Incubator and Development Division reach commercialization, they will be spun out as new public companies as it has done this twice already with Toggle3D.ai and ARway.ai. This strategic approach not only accelerates the path to market for these innovative solutions but also maximizes value for Nextech3D.ai shareholders. Following the successful precedents set by Arway.ai and Toggle3d.ai, shareholders in Nextech3D.ai will receive a stock dividend in these new entities, underscoring Nextech3D.ai's commitment to rewarding its investors and fostering a vibrant ecosystem of technology ventures.

CEO Evan Gappelberg stated, We are at the cusp of a new era in AI and 3D technology. With the launch of our AI Incubator and Development Division, and our unique approach to commercialization and value creation."

Hareesh Achi added, "It's an honor to join forces with Evan and the incredible team at Nextech3D.ai. Together, we're not just exploring the potential of AI; we're shaping its future, ensuring that the technologies we develop are used to make a meaningful difference in the world. The opportunity to directly reward our shareholders as we grow and spin out these innovations into new public companies is incredibly exciting."

The AI Incubator and Development Division is now actively working on several projects, investing heavily into the future of AI while also looking at possible acquisitions with more innovative solutions and updates expected to be unveiled in the coming months.

Read more here:
Nextech3D.ai Announces Formation of AI Incubator and AI Acquisition... - MarTech Outlook

Read More..

Unsupervised ensemble-based phenotyping enhances discoverability of genes related to left-ventricular morphology – Nature.com

In the following, we present our GWAS results. First, we investigate handcrafted phenotypes. Second, we examine unsupervised phenotypes obtained via shape PCA. Finally, we examine the results of our proposed UPE approach.

The loci were annotated with gene names on the basis of proximity to the lead single nucleotide polymorphism (SNP) if there was no additional causal evidence in the literature, or with nearby genes likely to mediate the association. For this, we used a diverse array of tools: the functional mapping and annotation (FUMA) web tool14, g:Profiler15, S-PrediXcan16 and the Ensembl Biomart database17. Among the candidate genes provided by these tools, a literature review was conducted to find evidence of an association with cardiovascular phenotypes, or experimental. Genes with asterisks were annotated solely on the basis of proximity and hence constitute totally novel findings.

We performed GWAS on traditional cardiac indices obtained using our segmentation approach. These indices were LVEDV, LV sphericity index at end diastole (LVEDSph), LV myocardial mass (LVM) and LV mass-to-volume ratio (LVMVR=LVM/LVEDV). Note that the LVEDSph as calculated here has not been investigated in previous GWAS (although a related phenotype, named LV internal dimensions was studied in an early GWAS of echocardiography-derived LV traits18). Details on how to compute this phenotype can be found in the Supplementary Information.

In the following, we discuss the associations found for each of these phenotypes. The Manhattan plots are shown in Extended Data Figs. 14.

For LVEDV, we discover nine independent associations. The association at intergenic SNP rs11153730 is probably related to PLN. This gene plays a crucial role in cardiomyocyte calcium handling by acting as a primary regulator of the SERCA protein (sarco- or endoplasmic reticulum Ca2+-ATPase), which transports calcium from the cytosol into the SR1 (ref. 19). Mutations in PLN have a well-established relationship with dilated cardiomyopathy (DCM)20. In ref. 4, PLN was found to be associated with LVEDV and LVESV. However, ref. 2 does not report this locus for the same phenotypes. The locus on chromosome 2 (with lead SNP rs2042995) is widely known to be associated with TTN. This gene encodes the protein titin, which is responsible for assembling myocyte sarcomere, and determines the stretching, contraction and passive stiffness of the myocardium21. This gene has been reported by refs. 2,4,11. rs375034445 lies within the body of BAG3; this is a well-known cardiac gene coding for a cellular protein that is predominantly expressed in skeletal and cardiac muscle, which plays a role in myocyte homeostasis and in the development of heart failure22; also, it shows a stronger association with LVESV and LV ejection fraction (LVEF), as found in previous studies2,4. The locus near the ATXN2 gene has previously been reported for LVEDV and stroke volume (SV)4. A candidate casual gene for this association is gene MYL2, the lead SNP (rs35350651) lies 558808 base pairs away from this genes transcription start site (TSS)23. The gene TMEM43 has been found in ref. 4 in association with LVESV and LVEF. Finally, gene MYH6 harbours SNP rs365990. This gene provides instructions for making a protein known as the cardiac -myosin heavy chain, which is expressed throughout the myocardium during early cardiac development24. Mutations in this gene, as well as the neighbouring MYH7 responsible for the -myosin heavy chain, have been linked to several pathologies: cardiomyopathies, arrhythmias and congenital heart disease (CHD). Two additional associations are located close to genes RRAS2 and ATG4D, respectively.

For LVEDSph, we find nine additional independent associations, apart from the PLN locus. rs35564079 is located 8,250bp upstream of the TSS of NKX2-5, in chromosome 5. This gene plays a crucial role in heart development; in particular, in the formation of the heart tube, which is a structure that will eventually give rise to the heart and great vessels. NKX2-5 helps determine the hearts position in the chest and also develops the heart valves and septa. Mutations in the NKX2-5 gene have been associated with several types of congenital heart defect, including atrial septal defects and atrioventricular block25. It has not been reported in refs. 2 or 4, but shows borderline significance with the fractal dimension of the LV trabeculae11. rs72007904 is located 300kb upstream of the TSS of the gene ABRA. ABRA codes for a cardiac and skeletal muscle-specific actin-binding protein located in the Z disc and M-line and binds with actin. Consistent with this, it is differentially expressed in cardiac tissues and skeletal muscle in the genotype-tissue expression (GTEx) data. ABRA has been associated with DCM in mice26. rs35001652 is close to KDM1A, a gene that codes for a histone demethylase involved in cardiac development, according to studies in mice27. rs463106 lies in the body of gene PRDM6. The mouse homologue of this gene, Prdm6, has been found to be important in early cardiac development28. An interesting association, with SNP rs162746, is close to gene EN1, however, we were not able to find a strong candidate gene in this region. Finally, rs573709385 lies in a gene desert in chromosome 2, the closest protein-coding genes are ACVR2A and ZEB2 (both at around 1.6Mb).

For LVM, four associations are found: rs4767239 is probably related to developmental gene TBX5 (T-box transcription factor 5), which has a known role in developing the heart and the limbs29. Through familial studies, mutations in this gene have been associated with Holt-Oram syndrome, a developmental disorder affecting the heart and upper limbs. In particular, there have been no recent reports on GWAS on LV phenotypes. The locus near the CENPW gene has a cardiac gene, HEY2, possibly causal for this association. HEY2 has been shown to suppress cardiac hypertrophy through an inhibitory interaction with GATA4, a transcription factor that plays a key role in cardiac development and hypertrophy30. HEY proteins are direct targets of Notch signalling and have been shown to regulate multiple key steps in cardiovascular development. Studies have found that the loss of HEY2 in mice leads to cardiac defects with high postnatal lethality31. This locus has also been reported as associated to right-ventricular phenotypes32. rs3740293 overlaps gene SYNPO2L, which is highly expressed in cardiac tissues (LV and atrial appendage) and skeletal muscle, making it a strong candidate gene. This SNP is also close to gene MYOZ1, which is also supported by our GWAS study (section on transcriptome-wide association studies, TWAS). Both genes have been previously proposed as candidates for cardiac phenotypes, in particular atrial fibrillation33,34. However, MYOZ1 shows very high expression only in the latter. Loss-of-function variants in this SYNPO2L have also been found causative of atrial fibrillation35, supporting this gene as a more likely candidate. rs73243622 is close to the candidate gene PPARGC1A. Finally, gene CDKN1A has been found in ref. 4 in association with LVESV and LVEF. Finally, for LVMVR, three new loci were found, apart from the PLN locus: rs2070458 close to SMARCB1 (in chromosome 22), rs17460016 in the FNDC3B locus (in chromosome 3) and rs12542527 (in chromosome 8). The last is an eQTL for the MTSS1 gene also linked to LV fractal dimension11.

The detailed summary statistics for the significant associations with handcrafted phenotypes are provided as Supplementary Data.

A shape PCA model was fit to our set of meshes (Methods). The effect on LV shape for the first 16 modes is shown in the Supplementary Fig. 7. GWAS was performed for these 16 modes and 18 independent loci were found with study-wide significance (P<3.1109). PC1, which is highly correlated with LVEDV, reconfirms the associations with TTN, MYL2 and MYH6. A new association, in chromosome 4, is an indel (chr4:120304290_GC_G) located 200kb downstream of MYOZ2. This gene codes for protein that functions by tethering calcineurin to alpha-actinin at Z-discs in muscle cells and inhibits the pathological cardiac hypertrophic response36. Another candidate gene in this locus is PDE5A. Indeed, some of the strongest associations overlap the body of this gene (although not the lead variant, which is the indel mentioned above). It has been shown that PDE5A is expressed in cardiac myocytes and may have pro-hypertrophic effects37.

PC2 is strongly linked with a new locus in chromosome 17, GOSR2. This component seems to be linked to LV conicity. Ref. 11 reports the GOSR2 locus as significantly associated with trabecular fractal dimension in slices 3 and 4, however, previous GWAS in global LV indices have not reported this locus. More broadly in the literature on genetics of cardiovascular phenotypes, it has been reported as associated to ascending aorta distensibility38, mitral valve geometry39 and CHD40.

PC3, highly correlated with LVEDSph, re-discovers the PLN and NKX2-5 loci. It also adds an association in chromosome 1, the SNP rs12142143, which lies within the ACTN2 gene. This gene codes for the Z disc protein -actinin-2. This locus has been reported for SV in ref. 4.

PC6 has hits in the TBX5 and NKX2-5 loci, with a new association near the NAV3 gene, that has been found to play a role in heart development in zebrafish41. PC7 is associated to a SNP near the TSS of PITX2 gene. It encodes for a transcription factor required for mammalian development, and disruption in its expression in humans causes CHD and is associated with atrial fibrillation. PC10 is linked to the PRDM6 locus (discussed before in connection with LVEDSph). PC11 is associated to SNPs rs59894072 (close to TBX3, a known cardiac gene42) and rs56229089. The second, in turn, is close (1Mb) to two possible candidate genes: KCNJ2, a potassium channel gene that is active in skeletal muscles and cardiac muscles43 and SOX9, a gene implicated in cardiac development44. The detailed summary statistics for the significant associations with shape PCs are provided in Supplementary Data.

CoMAs were trained on LV meshes at end diastole, using a range of network hyperparameters. The reconstruction performance for these models is shown in Supplementary Fig. 1.

GWAS was performed on all latent variables, for all training runs achieving a good reconstruction performance (Methods). A run is an instance of model training, defined by the choice of hyperparameters: in particular, random seeds controlling training and validation samples, weight initialization, network architecture and KullbackLeibler divergence weight. The number of such runs was R=36. The results obtained with nz=8 and nz=16 (8 and 16 latent variables, respectively) are reported, with a total number of 384 latent variables in the pooled representation. First, we examine the prevalence of significant GWAS loci found in all runs of our ensemble. To count the loci, we split the genome into approximately linkage disequilibrium-independent genomic regions45 and computed the number of loci below the usual genome-wide significance threshold of 5108 (see details in the Methods section); Table 1 shows the results.

We found 49 independent associations with study-wide significance. All of the previously discussed findings are recovered by UPE with study-wide significance, except the following loci: MTSS1, TBX3, PPARGC1A and FNDC3B (the last two show with suggestive significance in UPE). The summary statistics of the GWAS for the best latent variable of each of these 49 loci are displayed in Table 1. When a gene name is displayed in bold letters, it means that this locus was found only via the ensemble approach. Most loci have previous evidence supporting their plausible role in cardiac pathways. In addition, many of them are totally novel and represent interesting avenues for further research.

In what follows, we perform an in-depth analysis of our novel genetic findings in the light of recent literature.

We now describe loci that have not been linked to structural LV phenotypes in recent GWAS, but count with other types of evidence.

rs11706187 is probably linked to developmental gene SHOX2. The mouse homologue of SHOX2, Shox2, is essential to differentiate cardiac pacemaker cells by repressing Nkx2-5 (ref. 46). Whereas both TBX5 and NKX2-5 are highly expressed in adult cardiac tissues according to GTEx data, SHOX2 is not highly expressed in these tissues. A possible hypothesis is that rs11706187 regulates the expression of SHOX2 in developmental or pre-adult stages.

A particularly interesting association, with the SNP rs2245109, is located within the body of the STRN gene on chromosome 2 and is probably causally related to it: this gene encodes the protein striatin, which is expressed in cardiomyocytes and has been shown to interact with other proteins involved in the mechanism of myocardial function47. Mutations in this gene have been shown to lead to DCM in dogs48. In humans, there has been a recent GWAS on heart failure that reported this locus, but our study links it with cardiac morphology. Moreover, our estimated effect size is substantially higher; suggesting that this latent variable is an endophenotype closer to the underlying biology. This could provide insight to unravel the aetiology of a heterogeneous condition such as heart failure. The lead SNP has a high minor allele frequency (MAF) of 47.4%. This locus also contains eQTLs for this gene, as evidenced by TWAS (section TWAS). Something similar occurs with the RNF11 locus, although this does not reach genome-wide significance for heart failure (P=3.2106). The lead variant for this locus is an indel with low frequency (MAF 1.4%) and large estimated standardized effect size ((hat{beta }=)0.138). This locus has also been linked to the QRS (a combination of the Q, R and S waves) interval, although the causative gene is not clear49, some candidates being RNF11 itself, CDKN2C, C1orf185 and FAF1.

The SRL gene, which encodes the sarcalumenin protein, harbours the SNP rs889807. Sarcalumenin is a protein that binds Ca2+ located in the longitudinal sarcoplasmic reticulum of the heart. Its main function is to regulate Ca2+ reuptake in the sarcoplasmic reticulum by interacting with the cardiac sarco (endo)plasmic reticulum Ca2+-ATPase 2a (SERCA2a). According to GTEx data, this gene is highly expressed in adult cardiac tissue (both in the LV and atrial appendage tissues) and skeletal muscle.

Several associations lie near genes of the ADATMS (a disintegrin and metalloproteinase with thrombospondin motifs) family50: ADAMTS1 and ADAMTS5 (near rs2830977 on chromosome 21, with P=1.41010), ADAMTS6 (rs753963943 on chromosome 5, P=5.61011) and ADAMTS18 (chromosome 16, P=5.21013).

An association lies 260kb upstream of GATA6, a transcription factor that plays a critical role in the development of the heart. It has been found to regulate the hypertrophic response51. Sequence variants in this gene have been discovered to predispose for CHD phenotypes52,53.

rs12889267 lies 3,700kb upstream of the TSS of NDRG2. This gene has been demonstrated to play a role in protection against ischaemia and/or reperfusion injury, in a study in rats54.

One SNP overlaps KDM2A. As KDM1A, it is a histone demethylase gene. Although its link to the heart is less clear, there exists evidence from knockout studies in mice that supports its importance in embrionic development, including heart development55.

rs206524 is located within a gene for long non-coding RNA, LINC01254. A possible candidate protein-coding gene is NDUVF2, located 1.3Mb upstream of the SNP. According to the GTEx dataset, NDUFV2 is highly expressed in cardiac and skeletal muscle tissue.

rs12046416 is located 8,268bp upstream of the TSS of GJA5, a gene that is expressed in atrial myocytes and mediates the coordinated electrical activation of the atria56.

In addition to the loci with previous evidence discussed above, we report a number of novel genetic loci with P

In addition to genetic loci with P

A cluster of associations in chromosome 1 is located in a region that includes the S100 family of genes. In particular, the lead SNP in this region, rs985242, is located within the genes S100A1 and S100A13. The S100 is a family of low-weight Ca2+-binding EF-hand proteins, with 25 human genes identified.

The SNP rs28681517 lies within gene ADAMTSL3, whose associated protein has been shown to play a crucial role in maintaining cardiac structure and function in mice58.

SNP rs569550 lies 578,846 base pairs away from KCNQ1, which belongs to a large family of genes that provide instructions for making potassium channels. KCNQ1 encodes the alpha subunit of the potassium channel KvLQT1. Mutations in KCNQ1 are responsible for the long QT syndrome59.

Deletion 15:48690566_TC_T is a relatively common variant (MAF 14.4%), and is located 10kb downstream of the transcription end site of FBN1. Mutations in this gene are associated with Marfan syndrome, a genetic disorder that affects connective tissues in the body. It can have various manifestations, including cardiovascular complications.

rs9814240 is a coding variant in the LMCD1 gene. Mutations in this gene are causative of hypertrophic cardiomyopathy in mice60, however, no association had been found between variants in this gene and human cardiac phenotypes. Moreover, this gene has been found to interact with (the homologous of) GATA6 in mice61. GATA6 is located near one of the loci discovered with study-wide significance.

The effect of these loci on the LV morphology was evaluated by selecting the single phenotype with the strongest P value for the associated locus. To help characterize these latent variables, the Spearman correlation coefficient between the latter and the handcrafted LV indices were calculated and shown in Supplementary Table 4. We also examine the shapes of the average mesh within different ranges of quantiles for this latent variable, from 0 through 1. This is shown in Fig. 2, along with the associated Manhattan plots, for the loci PLN, TTN and STRN. The direction of effect is shown by indicating with arrows which allele favours which shape. We observe a very distinct effect on the morphology of each of these SNPs. While the PLN variant influences a latent variable that has a a smaller effect on LVEDV (Spearman r=0.722) and a strong link to LVEDSph (r=0.532), the best latent variable for TTN gene shows a greater correlation with LVEDV (r=0.910). Consistent with this, the GWAS on LVEDSph shows no significant signal for TTN, but a strong one for PLN (P=1020, Extended Data Fig. 2), which is also in line with a previous finding of ours10. Furthermore, these findings are in line with the effects of PC1 and PC3, where TTN and PLN loci are found, respectively.

ac, Manhattan plots for LV latent variables with best association for SNPs at the PLN (a), TTN (b) and STRN (c) loci. On top are shown the average meshes corresponding to the following range of quantiles, for each latent variable (from left to right): [0, 0.01], [0.095, 0.105], [0.495, 0.505], [0.895, 0.905] and [0.99, 1].

The SNP in the STRN gene is associated with a subtle phenotype that controls mitral orientation without a concomitant change in LV size (Fig. 2). This is consistent with the fact that it was not discovered in previous studies of structural LV phenotypes. Notably, this effect is consistent with the observed effect of PC4, for which this locus reaches genome-wide significance (see Supplementary Fig. 2 for the effect of PC4).

We performed TWAS using the S-PrediXcan tool16, to test the possibility of a mediating effect of gene expression and intron excision events on structural phenotypes. This tool is fed with models that impute gene expression and intron excision data on the basis of the genotype, which in turn were trained using data from the GTEx project, v.8 (ref. 62).

Our focus was on cardiovascular tissues, specifically the LV, atrial appendage and coronary, aortic and tibial arteries. To maintain statistical rigour, we applied a significance threshold of PGEx=2.2109, which adjusts for multiple comparisons (324 phenotypes and 68,919 tissuegene pairs). Similarly, for alternative splicing, the threshold was set at PAS=8.21010, considering the same multiple testing correction (187,535 being the number of intron-tissue pairs tested).

In the cardiac tissues (LV and atrial appendage), we identified genes located within loci of previously reported genes. In the LV, these included NKX2-5, STRN, SYNPO2L (FUT11, SEC24C and SYNPO2L itself), PLN, HEY2 (CENPW gene), TTN (FKBP7 gene), CENPV, GOSR2 (MAPT and GOSR2 itself) and FDPS (SCAMP3, ARHGEF2, RIT1, GOSR2, MAPT, HCN3, GBA, MSTO1, RUSC1, FUT11, SYT11, ADAM15 and FDPS itself). For the atrial appendage, the genes included PLN, STRN, NKX2-5, SYNPO2L and MYOZ1 within the SYNPO2L locus, as well as FKBP7 and SCAMP3. Many of these genes had been previously implicated on the basis of independent knowledge, bolstering the evidence for their potential causal roles. Notably, our analysis also revealed the direction of the effect on gene expression: higher PLN expression was associated with a more spherical LV morphology, while lower NKX2-5 expression was linked to the same phenotype (refer to Fig. 2b). Furthermore, an elevated STRN expression (in both cardiac tissues) was associated with a more horizontal mitral orientation (Fig. 2c). Detailed results for significant gene expression associations are provided as Supplementary Data.

In the case of arterial tissues, we found significant associations within various loci, such as the SYNPO2L locus (with the genes AGAP5, FUT11, SEC24C and ARHGAP27), FDPS (ARHGEF2, CLK2, FAM189B, GBA, GON4L, HCN3, NPR1 and SYT11), CENPW, TTN (PRKRA and FKBP7 genes), PLN (CEP85L and PLN), GOSR2 (WNT3, CRHR1, LRRC37A and MAPT), KDM2A, LINC01562, MYH6 (MYH6 and MYH7), RP11-383I23.2, RP11-574K11.29, SCAMP3, MYL2 (SH2B3 gene), SOST and TCF21.

Detailed results for intron excision events are provided in Supplementary Data.

We use the tool g:Profiler to find pathways for which our sets of genes were enriched. To define the gene sets, we selected a region of 100kb around each lead variant and chose the genes whose TSS was located within that window. Gene ontology terms belong to one of three different categories: molecular functions, cellular components and biological processes. Within the cellular component category, we have found a relevant enriched term, Sarcomere, comprising the following nine genes from our query: ACTN2, MYOZ1, SYNPO2L, BAG3, TNNT3, TNNI2, MYH6, MYH7, KY (P=9.2103). Within the biological process category, the terms Myofibril assembly, striated muscle cell development and sarcomere organization result enriched (P=1.2103, P=1.4103 and P=1.5103, respectively). Within the molecular function category, the term calcium-dependent protein binding is enriched (P=2.9108), although it is composed of nine members of the S100A family (which encompass a single locus), apart from SYT8 and TNNT3.

To detect pleiotropic effects, we performed a phenome-wide association study of the lead SNPs from Table 1. For this, we queried the Integrative Epidemiology Unit OpenGWAS Projects database. The results are included in the Supplementary Data File. We discuss briefly here some associations with cardiovascular phenotypes. A number of loci were associated to cardiac electrical phenotypes: CDKN1A, NDRG2, PLN, TBX5 and MYH6. The following loci were associated to pulse rate: SYNPO2L, NDRG2, MYH6, SRL, GOSR2, GATA6, ACTN2, KIAA1755, TMEM43, SLC27A6 and FNDC3B. The lead SNP at the PRDM6 locus was associated to heart rate recovery post exercise. The following loci were associated to blood pressure phenotypes (diastolic, systolic or hypertension): SYNPO2L, KCNQ1, MYL2, NDRG2, MYH6, SRL, GOSR2, GATA6, HSPB7, RNF11, EFEMP1, FNDC3B, NME9, PRDM6 and PLN. Finally, SYNPO2L, TBX5, MYH6, GOSR2, PITX2 and CDKN1A were associated to cardiac arrhytmias.

We set apart a subset of 5,470 UKBB participants of British ancestry for which the whole pipeline was run identically to the individuals from the discovery set. We report the detailed results in the Supplementary Material, including the estimated statistical power for each SNP on the basis of the effect size estimate (hat{beta }) from the discovery phase. Among the 49 study-wide significant loci, we report 28 that replicate with P<0.05 (whereas seven replicate with the more stringent Bonferroni threshold of P<0.05/49), as well as 47 loci for which the estimated direction of effect is consistent with that found in the discovery phase. For the suggestive associations, 11 loci replicated (out of 25) with the threshold of P<0.05, whereas 22 have a concordant direction of effect between the discovery and replication phases.

For comparison, we collected the GWAS summary statistics from previous studies on LV phenotypes, derived also from UKBB CMR images, namely refs. 2,4 and 11. We also include the results for LVESV, SV and LVEF from these studies. However, note that the unsupervised features studied in this work are static and were extracted using only the end-diastolic phase.

The comparison can be seen in Fig. 3. For each locus in Table 1 (which all pass the Bonferroni threshold), this figure displays the association P value found in previous GWAS and on our own GWAS of handcrafted phenotypes. Shades of red represent non-genome-wide significant associations, whereas shades of blue represent genome-wide significant ones and white corresponds to the PGW threshold. The second column represents the best P value across all traditional phenotypes for the loci given in the columns. Therefore, a shade of red in this column means that the locus is novel in the context of LV structural phenotypes.

The leftmost column corresponds to the best association found for that locus across the ensemble of phenotypes, whereas the second column corresponds to the best P value for that locus across the previous GWAS, where the P values are two-sided and derived from a linear association t-statistic (no adjustments were made for multiple comparisons). The white colour corresponds to the genome-wide significance threshold of 5108, whereas the shades of red and blue correspond to weaker and stronger associations, respectively. SV denotes stroke volume. LVEDVi, LVESVi and SVi denote the indexed versions of the phenotypes, that is, the phenotype divided by the participants body surface area. Finally, LVFD sn stands for LV trabecular fractal dimension measured at the nth slice of the LV longitudinal axis (for details, refer to the original publication in ref. 11). Grey squares (NA, not applicable) mean that the genetic variant was not tested in the corresponding study.

Source data

Originally posted here:
Unsupervised ensemble-based phenotyping enhances discoverability of genes related to left-ventricular morphology - Nature.com

Read More..

A comparative study of 11 non-linear regression models highlighting autoencoder, DBN, and SVR, enhanced by SHAP … – Nature.com

Result 1: Evaluation of the prediction accuracy of 11 AI models

Given the complex relationship between phenotype and genotype, genes can contribute positively or negatively to traits. Some genes have a significant impact, while others have a minor influence. So, we applied the non-linear regression algorithm in this study.

Considering the specific problem of predicting phenotypes and the characteristics of the SNP genotype dataset, we conducted a comparison using the most suitable non-linear regression machine learning and deep learning models. Seven machine learning models and four deep learning models were selected for this purpose. The seven machine learning models include SVR (Support Vector Regression), XGBoost (Extreme Gradient Boosting) regression, Random Forest regression, LightGBM regression, Gaussian Processes (GPS) regression, Decision Tree regression and Polynomial regression; The four deep learning models include Deep Belief Network (DBN) regression, Artificial Neural Networks (ANN) regression, Autoencoders regression, and Multi-Layer Perceptron (MLP) regression.

In this study, we collected phenotype data from 1918 soybean accessions and applied the corresponding SNP genotype data in our research. To address the large dataset size and redundancy of genotype data, we employed two steps. First, we used the one-hot encoder to convert the genotype data (ATCG nucleotide code) into an array. Then, Principal Component Analysis (PCA) was used to reduce the dimensionality of the data. Finally, the chosen models were applied using the respective algorithms.

GridsearchCV is a cross-validation procedure. To determine the optimal parameters, we employed the GridsearchCV method to fine-tune the hyperparameters and identify the best model for phenotype prediction. In order to evaluate the performance of the regression models, we utilized four evaluation metrics: R2 (R-squared), MAE (Mean Absolute Error), MSE (Mean Squared Error), and MAPE (Mean Absolute Percentage Error). These metrics were used to assess the prediction accuracy of each model.

Our results showed that, among seven machine learning models and four deep learning models, Polynomial regression exhibited the highest training performance, with an R2 value of 1.000, a MAE value of 0.00, an MSE value of 0.000, and an MAPE value of 0.000 during training, indicating a very close match between predicted and actual values. (Table 1). Among the seven machine learning models evaluated, lightGBM demonstrated the highest R2 value for the training set, achieving an impressive score of 0.967. Following closely is SVR, which obtained an R2 value of 0.926 for the training phase. Moving on to the test set, the SVR model performed the best, achieving the highest R2 value of 0.637, closely followed by Polynomial Regression with an R2 value of 0.614. Regarding the Mean Absolute Error (MAE) metric, the lightGBM model exhibited the lower value for the training set, registering a MAE of 0.068. Conversely, for the test set, the Polynomial Regression achieved the lowest MAE of 0.216, followed by the SVR model showcased the lower MAE value of 0.237. Considering the Mean Squared Error (MSE) metric, the lightGBM model demonstrated the lower value for the training set, yielding an MSE of 0.009. On the other hand, for the test set, the SVR model achieved the lowest MSE value, which amounted to 0.096, Followed by the Polynomial Regression which has MSE value of 0.102. Focusing on the Mean Absolute Percentage Error (MAPE), the Polynomial Regression model displayed the lowest value for the training set, and then the lightGBM model obtaining an MAPE of 0.025. In contrast, thePolynomial Regression and SVR model secured the lower MAPE for the test set, recording a value of 0.080 and 0.086 respectively (Table 1).

Among the four deep learning models, in training phase, ANN model got the highest R2 for train value of 0.995; and the lowest MAE for train value of 0.011, lowest MSE for train value of 0.001 and Lowest MAPE for train value of 0.004 (Table 1). when comparing with evaluation metrics R2, MAE and MSE in testing phase, the Autoencoder model got the best performance as mentioned above.

When comparing with other evaluation indicators, among all the models evaluated, the Autoencoder model had the highest R2 value for the test set, reaching an impressive 0.991. Additionally, the Autoencoder model obtained the lowest MAE value of.0.034 and the lowest MSE value of 0.002 during testing, indicating an excellent fit. (Table 1). Furthermore, the Autoencoder achieved the lowest MAPE value of 0.1011 during testing, indicating its good performance on unseen data.

Examining the test results, the correlation analysis reveals that R2_Autoencoder (0.991) outperforms R2_DBN (0.704), R2_SVR (0.637), and R2_Polynomial Regression (0.614). In the MAE analysis, MAE_Autoencoder (0.034) is lower than MAE_DBN (0.2.1), MAE_Polynomial Regression (0.216), and MAE_SVR (0.237). The MSE analysis shows that MSE_Autoencoder (0.002) is less than MSE_DBN (0.082), MSE_SVR (0.096), and MSE_Polynomial Regression (0.102). Regarding the MAPE analysis, MAPE_Autoencoder (0.011) is lower than MAPE_DBN (0.072), MAPE_Polynomial Regression (0.080), and MAPE_SVR (0.086).

In summary, based on our analysis of predictive model accuracy, the top four models are Autoencoder, DBN, SVR, and Polynomial Regression. This includes two machine learning models, SVR and Polynomial Regression, and two deep learning models, Autoencoder and DBN.

It should be noted that several articles have highlighted the drawbacks of percentage error metrics like MAPE. Caution is advocated by Stephan and Roland against relying on MAPE for the selection of the best forecasting method or the rewarding of accuracy, with an emphasis on the potential pitfalls associated with its minimization (Stephan and Roland, 2011)35. To further assess the performance of each model and gain a deeper understanding of the relative disparities between testing and training results, considering the magnitudes of the values being compared, we employed Relative Difference Analysis (RDA) on all four-evaluation metrics (Table 1). Our findings revealed that among the 11 models analyzed, Autoencoders demonstrated the most favorable performance, with a relative difference value of 0.001for R2, 0.014 for MAE, 0.046 for MSE, and 0.008 for MAPE. Additionally, the Decision Tree model achieved the lowest relative difference value for MAPE, with a value of 0.156. However, it also exhibited the highest R2 relative difference value at 1.035. On the other hand, the Polynomial Regression model displayed the highest relative difference values, with 2.000 for MAE, 2.000 for MSE, and 2.00 for MAPE (Table 1).

It has come to our attention that the R2 test score of the Multiple Linear Regression (MLR) model is significantly lower, approximately four times, than the R2 train score of the Autoencoder model. Additionally, the test loss in the MLR model is noticeably higher when compared to the train loss. The MAE for test score is at 0.331, which is a staggering 1.08E+14 times greater than the MAE for train of the MLR model. Additionally, the MSE for test of the MLR model stands at 1.26E+28 times higher than the MSE for train (Table 1). This observation strongly suggests a severe case of underfitting in the MLR model, as depicted in Table 1. Underfitting is distinct from overfitting, where the model may perform well on the training data but struggles to generalize its learning to the testing data. Underfitting becomes evident when the model's simplicity prevents it from establishing a meaningful relationship between the input and the output variables. The presence of underfitting in the MLR model signifies that the Linear model is too simplistic to be effectively utilized for phenotype prediction.

In order to further evaluate these 11 models, we plotted prediction accuracy evaluation based on Mean Absolute Error (MAE) (Fig.1), as well as overfitting evaluation based on Mean Squared Error (MSE) (Fig.2).

Histogram of Prediction Accuracy Evaluation of 11 Models by MAE Value. In this Figure, the histogram displays accuracy scores in the model evaluation using Mean Absolute Error (MAE). The blue dots represent the target values of the training data (y_train_pred), while the orange dots correspond to the target values of the testing data (y_test_pred). The X-axis represents the true values, and the Y-axis represents the prediction values.

Overfitting Evaluation of 11 Models Based on MSE Value. In this Figure, the histogram illustrates the evaluation of overfitting for each model. The blue line represents the Mean Squared Error (MSE) of the training data, while the orange line represents the MSE of the testing data. The Y-axis indicates the values of MSE, and the X-axis corresponds to different parameters for each model: For Decision tree, XGboost, and Random forest, the X-axis represents the max depth. For Gaussian process, the X-axis represents degrees of freedom. For SVR (Support Vector Regression), the X-axis represents the c-value. For lightGBM, the X-axis represents the number of iterations. For polynomial regression, the X-axis represents the polynomial degree. For DBN (Deep Belief Network) regression and Multilayer perception, the X-axis represents the hidden layer size. For Autoencoder and ANN (Artificial Neural Network) models, the X-axis represents the number of epochs. Each model's performance and overfitting tendencies can be observed and compared using these representations.

The probability plots of standardized residuals for each regression model provide a clear visual representation. The true values and predictions of the autoencoder model align well along the 45-degree line, with MAE of 0.03 for the training set and 0.03 for the test set. This demonstrates that the model's predictions adhere to the normality assumption. Similarly, the SVR model (MAE train=0.11 and MAE test=0.24), XGBoost model (MAE train=0.12 and MAE test=0.25), and DBN model (MAE train=0.03 and MAE test=0.02) also show good alignment between true values and predictions. On the other hand, the Multiplayer Perception model, Decision Tree model, Polynomial Regression model and MLR model exhibit a looser aggregation of true values and predictions, with data points scattered more loosely along the 45-degree line (Fig.1). The results of overfitting analysis indicate that SVR, lightGBM, Autoencoder, and ANN models fit both the training and test data exceptionally well, demonstrating a stable performance (Fig.2). While the testing loss of the MLP model shows significant fluctuations when the hidden layer size is below 400, it exhibits a robust fit for the training and test data when the hidden layer size exceeds 400. On the contrary, the Decision tree and DBN models demonstrate relatively poorer fits. As evident from the figures, the Decision tree model displays the least disparity between training and testing losses when the maximum depth (MAX Depth) is set to 5.0. Yet, when the depth is either below 5.0 or above 5.0, the gap between training and testing losses tends to widen. Regarding the DBN model, a relatively stable gap between training and testing losses is maintained for hidden layer sizes below 100. However, when the hidden layer size exceeds 100, the gap gradually increases. Similarly, the Polynomial regression model performs well when the polynomial degree is below 7. However, when the degree surpasses 9, there is a sharp increase in the gap between the training and testing losses (Fig.2). Both the Random forest and Gaussian process models exhibit a growing gap between training and testing losses with an increase in the maximum depth or the degree of freedom (degree of freedom) (Fig.2).

In summary, based on our comprehensive analysis, it is evident that Autoencoder, SVR, and ANN outperform the other models in relative terms. These models are suitable for genotype to phenotype prediction and minor QTL mapping. It could be the powerful tools in AI assisted breeding practice.

Our objective is to discover the most effective artificial intelligence model and utilize feature selection techniques to pinpoint genes responsible for specific physiological activities in plants. These identified genes will aid in precise phenotype prediction and gene function mining. To ensure the model's reliability, efficiency, low computational requirements, versatility, and openness, this study employs the Support Vector Regression (SVR) model as an illustrative example. We assess four distinct feature selection algorithms: Variable Ranking, Permutation, SHAP, and Correlation Matrix. Apart from the feature importance data, the Correlation Matrix method also provides valuable insights. A heatmap is employed to visualize the strength of correlations. In Fig.3, we present the heatmap showcasing the top 100 features identified through the Correlation Matrix analysis based on the SVR model (Fig.3). Additionally, the SHAP output plot offers a concise representation of the distribution and variability of SHAP values for each feature. Figure4 illustrates the summary beeswarm plot of the top 20 features derived from our SHAP importance analysis based on the SVR model. This plot effectively captures the relative effect of all the features in the entire dataset (Fig.4).

Correlogram of Top 100 Features (SNP) Identified in SVR Correlation Analysis. The figure displays a heatmap representing the correlations between the top 100 features (Single Nucleotide PolymorphismsSNP) identified in the SVR (Support Vector Regression) correlation analysis. The heatmap uses varying shades of the color gray, with higher values indicating stronger correlations between the variables. This visualization allows for a clear and visual assessment of the interrelationships among the features, providing valuable insights into their associations and potential implications in the study.

Summary Beeswarm Plot of Top 20 Features from SHAP Importance Analysis based on SVR Model. This figure presents a beeswarm plot summarizing the top 20 features derived from our SHAP (SHapley Additive exPlanations) importance analysis using the SVR (Support Vector Regression) model. The plot visually captures the relative effect of each feature across the entire dataset, allowing for a comprehensive understanding of their respective influences. The beeswarm plot provides an intuitive representation of the feature importances, aiding in the identification of key contributors to the model's predictions and facilitating insightful data-driven decisions.

We ranked all SNPs based on the absolute values of feature importance obtained from four feature selection methods respectively (see Supplementary 1). Considering that the ranking results do not follow a normal distribution and the assumptions of equal variances, we conducted a significance analysis of the differences in these rankings using the Wilcoxon signed-rank test, instead of the paired t-test.

Our results showed that the difference between Variable ranking and Permutation ranking is significant at P-value 0.05 level. The difference between Variable ranking and ranking of Correlation Matrix or SHAP were not significant. The difference between Permutation ranking and ranking of Correlation Matrix or SHAP were not significant. The difference between Correlation Matrix ranking and SHAP ranking was not significant also (Table 2.).

Compare to the importance results of other three methods, SHAP importance provide very rich information of negative contribution genes (Supplementary 1). Understanding the positive and negative contributions is vital for studying the gene's function and its role in plant physiological activities. Consequently, in the subsequent biological analysis, we made use of the SHAP importance results from our research.

By employing the basic local alignment search tool (BLAST), we conducted a comparative analysis of the sequences associated with 1033 single nucleotide polymorphisms (SNPs) against the annotated genes available in the soybase database (https://www.soybase.org/). Among these SNPs, 253 displayed a perfect match with their corresponding genes (refer to Supplementary 2). Subsequently, we performed a Gene Ontology (GO) analysis on these 111 genes and mapped their positions to the chromosomes of soybeans, as illustrated in Fig.5.

Whole Genome View of 111 Identified Genes. The figure presents a visual representation of identified genes, where each red dot represents a corresponding gene from the BLAST (Basic Local Alignment Search Tool) hit. The genes displayed in the plot are related to soybean branching. This comprehensive genome view provides valuable insights into the spatial distribution and clustering patterns of the branching-related genes, aiding in the exploration and understanding of their potential functional significance.

We conducted GO enrichment analysis on these 111 genes from three aspects: molecular function, cellular components, and biological process. Our analysis results revealed that the GO terms related to Biological Processes could be clustered into seven categories, with a total occurrence of 31 genes. The most prominent category was "signal transduction" (11 out of 31), followed by "translation" and "lipid metabolic process," each accounting for six out of 31 genesrespectively. Regarding Molecular Function, the GO terms could be grouped into 13 categories, with a total of 157 gene occurrences. The most prevalent category was "protein binding" (31 out of 157), followed by "transferase activity" (22 out of 157), and "kinase activity" (20 out of 157). Concerning Cellular Components, the GO terms could be classified into 21 categories, with a total of 380 gene occurrences. The most significant category was the "plasma memberane" (56 out of 380), followed by "cytoplasm" (42 out of 380), and "extracellular region" (42 out of 380). For detailed results, please refer to Fig.6 and Supplement 3.

Analysis of GO ontologies distribution. The figure displays three pie charts representing the distribution of three kinds of Gene Ontology (GO) ontologies, namely Cellular Component, Molecular Function, and Biological Process. Each pie chart is color-coded to distinguish different types of GO, and the size of each segment represents the proportion of that specific GO type within its respective ontology category. The accompanying number table provides the count of genes associated with each GO type, followed by the ID and category of the corresponding GO term. This analysis provides a comprehensive overview of the functional annotations of the genes in the study, highlighting their involvement in various cellular components, molecular functions, and biological processes.

Furthermore, we performed Gene Ontology enrichment analysis using the agriGO database. The outcomes revealed the functional distribution of 111 genes associated with biological processes (Fig.7). Notably, these processes exhibited a significant level (level 19) of overall metabolic activities. We observed a negative regulation between multicellular organismal processes and cell recognition. Additionally, a complex interplay of negative and positive regulations among reproduction-related processes, including reproductive process, pollination, pollen-pistil interaction, and recognition of pollen were detected (Fig.7).

GO Term Enrichment Analysis of 244 Genes using AgriGo Database Corresponding to Biological Function. The figure presents the results of GO term enrichment analysis performed on the 244 genes using the AgriGo database, focusing on their biological functions. The color shading in the illustration ranges from red to yellow, representing the significance levels of the enriched GO terms, with red indicating strong significance and yellow indicating weaker significance.Furthermore, different arrow types are employed to indicate the regulation relationships between the enriched GO terms and the genes. For instance, a green arrow signifies negative regulation, while other arrow types correspond to various regulation types.This analysis provides valuable insights into the functional annotations and regulatory relationships of the studied genes, shedding light on their roles and potential biological implications in the context of the AgriGo database.

Read more here:
A comparative study of 11 non-linear regression models highlighting autoencoder, DBN, and SVR, enhanced by SHAP ... - Nature.com

Read More..

The Adams administration quietly hired its first AI czar. Who is he? – City & State New York

New York City has quietly filled the role of director of artificial intelligence and machine learning, City & State has learned. In mid-January, Jiahao Chen, a former director of AI research at JPMorgan Chase and the founder of independent consulting company Responsible AI LLC, took on the role, which has been described by the citys Office of Technology and Innovation as spearheading the citys comprehensive AI strategy.

Despite Mayor Eric Adams administration publicizing the position last January, Chens hiring nearly a year later came without any fanfare or even an announcement. The first mention of Chen as director of AI came in a press release sent out by the Office of Technology and Innovation on Thursday morning, announcing next steps in the citys AI Action Plan. OTI Director of AI and Machine Learning Jiahao Chen will manage implementation of the Action Plan, the press release noted.

New York City previously had an AI director under former Mayor Bill de Blasios administration. Neal Parikh served as the citys director of AI under the office of former Chief Technology Officer John Paul Farmer, which released a citywide AI strategy in 2021. Under de Blasio, the city also had an algorithms management and policy officer to guide the city in the development, responsible use and assessment of algorithmic tools, which can include AI and machine learning. The old CTOs office and the work of the algorithms officer was consolidated along with the citys other technology-related offices into the new Office of Technology and Innovation at the outset of the Adams administration.

The Adams administration has referred to its own director of AI and machine learning as a new role, however, and has suggested that the position will be more empowered, in part because it is under the larger, centralized Office of Technology and Innovation. According to the job posting last January, which noted a $75,000 to $140,000 pay range, the director will be responsible for helping agencies use AI and machine learning tools responsibly, consulting with agencies on questions about AI use and governance, and serving as a subject matter expert on citywide policy and planning, among other things. How the role will actually work in practice remains to be seen.

The Adams administrations AI action plan was published in October, and isa 37-point road map aimed at helping the city responsibly harness the power of AI for good. On Thursday, the Office of Technology and Innovation announced the first update on the action plan, naming members of an advisory network that will consult on the citys work. That list includes former City Council Member Marjorie Velzquez, who is now vice president of policy at Tech:NYC. The office also released a set of AI principles and definitions, and guidance on generative AI.

OTI spokesperson Ray Legendre said that an offer for the position of director of AI was extended to Chen before the citys hiring freeze began last October. The office did not explicitly address why Chens hiring wasnt announced when he started the role. Over the past two months, Jiahao has been a key part of our ongoing efforts to implement the AI Action Plan, Legendre wrote in an email. Our focus at OTI over the past few months has been on making progress on the Action Plan which is what we announced today.

According to the website for Responsible AI LLC, Chens independent consulting company, Chens resume includes stints in academia as well as the private sector, including as a senior manager of data science at Capital One, and as director of AI research at JPMorgan Chase.

After City & State inquired about Chens role, Chen confirmed it on X, writing I can finally talk about my new job!

See more here:
The Adams administration quietly hired its first AI czar. Who is he? - City & State New York

Read More..