Page 1,665«..1020..1,6641,6651,6661,667..1,6701,680..»

Scientists Identify Thousands of New Cosmic Objects Using Machine Learning – SciTechDaily

Application of machine learning techniques to large astronomy data sets can discover thousands of cosmic objects of various classes. Credit: Shivam Kumaran

A team of scientists from the Tata Institute of Fundamental Research in Mumbai and the Indian Institute of Space Science and Technology in Thiruvananthapuram, consisting of Prof. Sudip Bhattacharyya and Mr. Shivam Kumaran, along with Prof. Samir Mandal and Prof. Deepak Mishra, have utilized machine learning techniques to identify the nature of thousands of new celestial objects in X-ray wavelengths. Machine learning is a branch of artificial intelligence.

Astronomy is undergoing a transformation as vast amounts of astronomical data from millions of celestial objects become readily accessible. This is due to large-scale surveys and meticulous observations utilizing top-notch astronomical observatories, combined with a policy of open data availability.

Needless to say that these data have great potential for many discoveries and a new understanding of the universe. However, it is not practical to explore the data from all these objects manually, and automated machine learning techniques are essential to extract information from these data. But the application of such techniques to astronomical data is still very limited and in the preliminary stage.

In this background, the TIFR-IIST team applied machine learning techniques to hundreds of thousands of cosmic objects observed in X-rays with USAs Chandra space observatory. This demonstrated how a new and topical technological progress could help and revolutionize basic and fundamental scientific research. The team applied these techniques to about 277000 X-ray objects, the nature of most of which were unknown.

A classification of the nature of unknown objects is equivalent to the discovery of objects of specific classes. Thus, this research led to a reliable discovery of many thousands of cosmic objects of classes, such as black holes, neutron stars, white dwarfs, stars, etc., which opened up an enormous opportunity for the astronomy community for further detailed studies of many interesting new objects.

This collaborative research has also been important to establish a state-of-the-art capacity to apply new machine-learning techniques to fundamental research in astronomy, which will be crucial to scientifically utilize the data from current and upcoming observatories.

Reference: Automated classification of Chandra X-ray point sources using machine learning methods by Shivam Kumaran, Samir Mandal, Sudip Bhattacharyya and Deepak Mishra, 9 February 2023, Monthly Notices of the Royal Astronomical Society.DOI: 10.1093/mnras/stad414

See the original post:
Scientists Identify Thousands of New Cosmic Objects Using Machine Learning - SciTechDaily

Read More..

Researchers From Tsinghua University Introduce A Novel Machine Learning Algorithm Under The Meta-Learning Paradigm – MarkTechPost

Recent achievements in supervised tasks of deep learning can be attributed to the availability of large amounts of labeled training data. Yet it takes a lot of effort and money to collect accurate labels. In many practical contexts, only a small fraction of the training data have labels attached. Semi-supervised learning (SSL) aims to boost model performance using labeled and unlabeled input. Many effective SSL approaches, when applied to deep learning, undertake unsupervised consistency regularisation to use unlabeled data.

State-of-the-art consistency-based algorithms typically introduce several configurable hyper-parameters, even though they attain excellent performance. For optimal algorithm performance, it is common practice to tune these hyper-parameters to optimal values. Unfortunately, hyper-parameter searching is often unreliable in many real-world SSL scenarios, such as medical image processing, hyper-spectral image classification, network traffic recognition, and document recognition. This is because the annotated data are scarce, leading to high variance when cross-validation is adopted. Having algorithm performance sensitive to hyper-parameter values makes this issue even more pressing. Moreover, the computational cost may become unmanageable for cutting-edge deep learning algorithms as the search space grows exponentially concerning the number of hyper-parameters.

Researchers from Tsinghua University introduced a meta-learning-based SSL algorithm called Meta-Semi to leverage the labeled data more. Meta-Semi achieves outstanding performance in many scenarios by adjusting just one more hyper-parameter.

The team was inspired by the realization that the network may be trained successfully using the appropriately pseudo-labeled unannotated examples. Specifically, during the online training phase, they produce pseudo-soft labels for the unlabeled data based on the network predictions. Next, they remove the samples with unreliable or incorrect pseudo labels and use the remaining data to train the model. This work shows that the distribution of correctly pseudo-labeled data should be comparable to that of the labeled data. If the network is trained with the former, the final loss on the latter should also be minimized.

They defined the meta-reweighting objective to minimize the final loss on the labeled data by selecting the most appropriate weights (weights throughout the paper always refer to the coefficients used to reweight each unlabeled sample rather than referring to the parameters of neural networks). The researchers encountered computing difficulties when tackling this problem using optimization algorithms.

For this reason, they suggest an approximation formulation from which a closed-form solution can be derived. Theoretically, they demonstrate that each training iteration only needs a single meta gradient step to achieve the approximate solutions.

In conclusion, they suggest a dynamic weighting approach to reweight previously pseudo-labeled samples with 0-1 weights. The results show that this approach eventually reaches the stationary point of the supervised loss function. In popular image classification benchmarks (CIFAR-10, CIFAR-100, SVHN, and STL-10), the proposed technique has been shown to perform better than state-of-the-art deep networks. For the difficult CIFAR-100 and STL-10 SSL tasks, Meta-Semi gets much higher performance than state-of-the-art SSL algorithms like ICT and MixMatch and obtains somewhat better performance than them on CIFAR-10. Moreover, Meta-Semi is a useful addition to consistency-based approaches; incorporating consistency regularisation into the algorithm further boosts performance.

According to the researchers, Meta-Semi requires a little more time to train is a drawback. They plan to look into this issue in the future.

Check out thePaperandReference Article.All Credit For This Research Goes To the Researchers on This Project. Also,dont forget to joinour 15k+ ML SubReddit,Discord Channel,andEmail Newsletter, where we share the latest AI research news, cool AI projects, and more.

Tanushree Shenwai is a consulting intern at MarktechPost. She is currently pursuing her B.Tech from the Indian Institute of Technology(IIT), Bhubaneswar. She is a Data Science enthusiast and has a keen interest in the scope of application of artificial intelligence in various fields. She is passionate about exploring the new advancements in technologies and their real-life application.

Read the original post:
Researchers From Tsinghua University Introduce A Novel Machine Learning Algorithm Under The Meta-Learning Paradigm - MarkTechPost

Read More..

Weights & Biases simplifies machine learning production and … – SiliconANGLE News

Machine learning development startup Weights & Biases Inc., whose software is used by the likes of OpenAI LLC and Nvidia Corp. to develop new artificial intelligence models, announced two major enhancements to its platform today.

Weights and Biases has created a platform for teams to build and collaborate on machine learning models and operations, or MLOps. The platform enables teams to keep track of their machine learning experiments. It also provides tools for evaluating the performance of different machine learning models, dataset versioning and pipeline management.

The W&B platform is designed to improve the efficiency of the trial-and-error process through which AI software is developed. According to the startup, it helps by increasing developer productivity, as it solves one of the main challenges of AI initiatives: organizing and processing project data.

The new additions to its platform were announced at Fully Connected, W&Bs inaugural user conference, and include W&B Launch and W&B Models.

Available in public preview starting today, W&B Launch provides an easier way for users to package code automatically and launch a new machine learning job on any target environment. In this way, machine learning practitioners gain easier access to the compute resources they need, while simplifying infrastructure complexity. In addition, W&B Launch makes it easier for teams to reproduce runs and scale up and scale out those activities.

Orlando Avila-Garcia, AI principal researcher at ARQUIMEA Research Center, said W&B Launch will make it easier for his organization to accelerate research on a variety of deep learning techniques, including neural radiance fields and graph neural networks. Abstracting away the complexity of using the infrastructure for our researchers is very beneficial to our overall team, he said. Launch greatly simplifies our work optimizing, experimenting with and benchmarking ML methods, letting us focus on reliability and reproducibility of results.

As for W&B Models, this is generally available now and provides a more scalable way for teams to govern machine learning model lifecycles in a centralized repository, while allowing for cross-functional discovery and collaboration. With its reproducibility and lineage tracking functionality that makes it possible for users to track exactly when a model was moved from staging to production, W&B Models can help teams to maintain higher-quality models over time.

Andy Thurai, vice president and principal analyst of Constellation Research Inc., told SiliconANGLE that Weights & Biases is one of the older players in the MLOps segment, and its especially strong inmodel creation, data set versioning and experiment tracking. He added that W&B Models looks like it will be a good addition to the companys solution set. It offers ML model registry and model governance, allowing users to collaborate more easily by making models discoverable across large enterprises, he said. This, combined with model lineage tracking, enables users to track models in the ML pipeline from inception to production.

Were confident that W&B users will quickly see the benefits of Launch and Models in accelerating model training, utilizing compute resources efficiently, managing models with more confidence, and having a more cohesive end-to-end ML workflow, said W&B Vice President of Product Phil Gurbacki.

The platform enhancements follow what W&B says has been a very promising 12 months, during which it has enjoyed significant traction and momentum. The company, which last raised $135 million at a $1 billion valuation in October 2021, claims to have doubled its employee base in the last year, opening new offices in Berlin, London and Tokyo to help boost its international presence.

See more here:
Weights & Biases simplifies machine learning production and ... - SiliconANGLE News

Read More..

Lemmatization in NLP and Machine Learning – Built In

Lemmatization is one of the most common text pre-processing techniques used in natural language processing (NLP) and machine learning in general. Lemmatization is not that much different than the stemming of words in NLP. In both stemming and lemmatization, we try to reduce a given word to its root word. The root word is called a stem in the stemming process, and its called a lemma in the lemmatization process. But there are a few more differences to the two than that. Lets see what those are.

Lemmatization is a text pre-processing technique used in natural language processing (NLP) models to break a word down to its root meaning to identify similarities. For example, a lemmatization algorithm would reduce the word better to its root word, or lemme, good.

In stemming, a part of the word is just chopped off at the tail end to arrive at the stem of the word. There are different algorithms used to find out how many characters have to be chopped off, but the algorithms dont actually know the meaning of the word in the language it belongs to. In lemmatization, the algorithms do have this knowledge. In fact, you can even say that these algorithms refer to a dictionary to understand the meaning of the word before reducing it to its root word, or lemma.

So, a lemmatization algorithm would know that the word better is derived from the word good, and hence, the lemme is good. But a stemming algorithm wouldnt be able to do the same. There could be over-stemming or under-stemming, and the word better could be reduced to either bet, or bett, or just retained as better. But there is no way in stemming that can reduce better to its root word good. This is the difference between stemming and lemmatization.

More on Machine Learning: An Introduction to Classification in Machine Learning

As you can probably tell by now, the obvious advantage of lemmatization is that it is more accurate than stemming. So, if youre dealing with an NLP application such as a chat bot or a virtual assistant, where understanding the meaning of the dialogue is crucial, lemmatization would be useful. But this accuracy comes at a cost.

Because lemmatization involves deriving the meaning of a word from something like a dictionary, its very time consuming. So most lemmatization algorithms are slower compared to their stemming counterparts. There is also a computation overhead for lemmatization, however, in most machine learning problems, computational resources are rarely a cause of concern.

More on Machine Learning: Multiclass Classification With an Imbalanced Data Set

Well, I cant answer that question. Lemmatization and stemming are both much more complex than what Ive made them appear here. There are a lot more things to consider about both the approaches before making a decision. But Ive rarely seen any significant improvement in efficiency and accuracy of a product that uses lemmatization over stemming. In most cases, at least according to my knowledge, the overhead that lemmatization demands is not justified. So, it depends on the project in question. But I want to put out a disclaimer here. Most of the work I have done in NLP is for text classification, and thats where I havent seen a significant difference. There are applications where the overhead of lemmatization is perfectly justified, and in fact, lemmatization would be a necessity.

Read the original here:
Lemmatization in NLP and Machine Learning - Built In

Read More..

Startup partners to bring machine learning, facial recognition to IoT … – Biometric Update

A partnership has been formed between Useful Sensors, which supports the addition of AI interfaces to consumer electronics, and OKdo, a subsidiary of RS Group which makes and sells single board computers, to increase the availability of machine learning capabilities like biometrics, presence detection, hand gesture recognition and voice interaction.

The partners plan to bring small, low-cost machine learning hardware modules to market through the global manufacturing and distribution deal.

Useful Sensors makes a coin-sized module with a camera and a microcontroller that can perform facial recognition, according to the announcement. Other applications for the companys person sensor technology include person detection, QR code scanning, and speech recognition.

Were excited to be working with OKdo, says Pete Warden, CEO and founder of Useful Sensors. The companys deep expertise in building products that solve real problems for their customers is going to improve the solutions were creating and help us come up with entirely new ways of helping system builders.

Useful Sensors was formed a year ago by Warden and CTO Manjunath Kudlur, who are both founding members of Googles TensorFlow open-source machine learning framework.

The companies are carrying out live demonstrations of their joint technology at Embedded World in Nuremberg, Germany this week.

biometric sensors | biometrics | facial recognition | machine learning | microcontroller | OKdo | person detection | research and development | Useful Sensors

See the original post:
Startup partners to bring machine learning, facial recognition to IoT ... - Biometric Update

Read More..

‘Meta-Semi’ machine learning approach outperforms state-of-the-art algorithms in deep learning tasks – Tech Xplore

by Tsinghua University Press

Meta-Semi trains deep networks using pseudo-labeled samples whose gradient directions are similar to labeled samples. Algorithm 1 shows the Meta Semi pseudo code. The Meta-Semi algorithm outperforms state-of-the-art semi-supervised learning algorithms. Credit: CAAI Artificial Intelligence Research, Tsinghua University Press

Deep learning based semi-supervised learning algorithms have shown promising results in recent years. However, they are not yet practical in real semi-supervised learning scenarios, such as medical image processing, hyper-spectral image classification, network traffic recognition, and document recognition.

In these types of scenarios, the labeled data is scarce for hyper-parameter search, because they introduce multiple tunable hyper-parameters. A research team has proposed a novel meta-learning based semi-supervised learning algorithm called Meta-Semi, that requires tuning only one additional hyper-parameter. Their Meta-Semi approach outperforms state-of-the-art semi-supervised learning algorithms.

The team published their work in the journal CAAI Artificial Intelligence Research.

Deep learning, a machine learning technique where computers learn by example, is showing success in supervised tasks. However, the process of data labeling, where the raw data is identified and labeled, is time-consuming and costly. Deep learning in supervised tasks can be successful when there is plenty of annotated training data available. Yet in many real-world applications, only a small subset of all the available training data are associated with labels.

"The recent success of deep learning in supervised tasks is fueled by abundant annotated training data," said Gao Huang, associate professor with the Department of Automation at Tsinghua University. However, the time-consuming, costly collection of precise labels is a challenge researchers have to overcome. "Meta-semi, as a state-of-the-art semi-supervised learning approach, can effectively train deep models with a small number of labeled samples," said Huang.

With the research team's Meta-Semi classification algorithm, they efficiently exploit the labeled data, while requiring only one additional hyper-parameter to achieve impressive performance under various conditions. In machine learning, a hyper-parameter is a parameter whose value can be used to direct the learning process.

"Most deep learning based semi-supervised learning algorithms introduce multiple tunable hyper-parameters, making them less practical in real semi-supervised learning scenarios where the labeled data is scarce for extensive hyper-parameter search," said Huang.

The team developed their algorithm working from the assumption that the network could be trained effectively with the correctly pseudo-labeled unannotated samples. First they generated soft pseudo labels for the unlabeled data online during the training process based on the network predictions.

Then they filtered out the samples whose pseudo labels were incorrect or unreliable and trained the model using the remaining data with relatively reliable pseudo labels. Their process naturally yielded a meta-learning formulation where the correctly pseudo-labeled data had a similar distribution to the labeled data. In their process, if the network is trained with the pseudo-labeled data, the final loss on the labeled data should be minimized as well.

The team's Meta-Semi algorithm achieved competitive performance under various conditions of semi-supervised learning. "Empirically, Meta-Semi outperforms state-of-the-art semi-supervised learning algorithms significantly on the challenging semi-supervised CIFAR-100 and STL-10 tasks, and achieves competitive performance on CIFAR-10 and SVHN," said Huang.

CIFAR-10, STL-10, and SVHN are datasets, or collections of images, that are frequently used in training machine learning algorithms. "We also show theoretically that Meta-Semi converges to the stationary point of the loss function on labeled data under mild conditions," said Huang. Compared to existing deep semi-supervised learning algorithms, Meta-Semi requires much less effort for tuning hyper-parameters, but achieves state-of-the-art performance on the four competitive datasets.

Looking ahead to future work, the research team's aim is to develop an effective, practical and robust semi-supervised learning algorithm. "The algorithm should require a minimal number of data annotations, minimal efforts of hyper-parameter tuning, and a minimized training time. To attain this goal, our future work may focus on reducing the training cost of Meta-Semi," said Huang.

More information: Yulin Wang et al, Meta-Semi: A Meta-Learning Approach for Semi-Supervised Learning, CAAI Artificial Intelligence Research (2023). DOI: 10.26599/AIR.2022.9150011

Provided by Tsinghua University Press

See the original post:
'Meta-Semi' machine learning approach outperforms state-of-the-art algorithms in deep learning tasks - Tech Xplore

Read More..

Machine Learning Program Sorts Plastic With High Success Rate – waste360

Studies show that just about 5 percent of all plastic that people want to recycle make it through the process and is renewed. Big factors for this low number can be attributed contaminated materials, water requirements, and discarded waste along with that drastic 263 percent increase in Americans plastic waste use since 1980.

A new machine learning model has been developed by a team at University College London that is able to isolate compostable and biodegradable plastics and help to improve recycling efficiency and accuracy.

Using data based on a classification system using hyperspectral imaging, a machine learning program has been able to observe and sort individual pieces of plastic waste.

Depending on the size of the material, the program has been able to achieve perfect accuracy in sorting. Plastic materials larger than 10mm by 10mm were sorted with 100 percent accuracy, while smaller material rates were less than perfect.

Read the full article here.

Read more from the original source:
Machine Learning Program Sorts Plastic With High Success Rate - waste360

Read More..

How machine learning is used in portfolio management? – Rebellion Research

How machine learning is used in portfolio management?

Artificial Intelligence & Machine Learning

Introduction

Modern Portfolio Theory was first introduced by Harry Markowitz in 1952. The groundbreaking investment theory demonstrated that the performance of an individual stock is not as important as the performance of an entire portfolio. The interpretations of his article on Portfolio Selection [1], in The Journal of Finance, taught investors that risk, and not the best price, should be the cornerstone of a portfolio [2]. Once an investors risk tolerance has been established, building a portfolio around it is streamlined.

The goal of portfolio management is to choose the best strategy to maximize returns on investment, ensure portfolio flexibility or optimize risk. Attaining such objectives involve the optimal allocation of investment funds, optimization of risk factors, considering financial ratios (such as the Sharpe Ratio) and much more. [3]

Such models help portfolio managers with trade execution, data parsing, idea generation & pattern recognition, alpha factor design, asset allocation, position sizing, strategy testing, and eliminates human biases.

Our goal of this project is to identify an efficient Machine Learning architecture that will replicate the stock basketstrend, almost perfectly. The selected ML model will have superior prediction capabilities and will adapt to the directional changes in prices accurately. Thus, the ML model will help us predict the expected stock returns. Additionally, using the past data we can calculate the standard deviations and gauge the risk.

Now, these stocks will be parsed through various portfolio optimization techniques. To identify the correct optimization technique, in-depth research on various techniques will be carried out. This will produce an optimized portfolio with weightage of each stock.

Literature Review & Model Implementation

1) Meanvariance portfolio optimization using machine learning-based stock price prediction (Wei Chen, Haoyu Zhang, Mukesh Kumar Mehlawat, Lifen Jia) [5]

The paper presents a hybrid model based on machine learning for stock prediction and mean variance (MV) model for portfolio selection. The model follows two stages:

Stock Prediction: Combines eXtreme Gradient Boosting (XGBoost) with an improved firefly algorithm (IFA). XGBoost is an improved gradient boosted decision tree and is composed of multiple decision trees. It is, therefore, a suitable classifier for the financial markets. The accuracy of prediction is calculated using mean absolute percentage error (MAPE), mean square error (MSE), mean absolute error (MAE), and root mean square error (RMSE).

Portfolio Selection: Stocks with higher potential returns are selected according to each stocks predicted prices, and the MV model is employed for allocating the investment proportion of the portfolio. The portfolio is evaluated on its annualized mean return, annualized standard deviation, annualized Sharpe ratio, and annualized Sortino ratio (variation of Sharpe ratio considering the standard deviation of only negative portfolio returns instead of the total standard deviation.)

The paper concludes that ML portfolio management pricing methods outperform the benchmarks concerning accuracy, potency, and efficiency.

Methodology

We predicted the stock price of three companies using ARIMA, LSTM-RNN, and Transformer architectures. The accuracy of prediction is calculated using root mean square error (RMSE), mean absolute error (MAE), mean absolute percentage error (MAPE), and Profit & Loss (PNL) rubrics.

a) Terminology

ARIMA[6][7] : AutoRegressive Integrated Moving Average is a time series forecasting model that incorporates autocorrelation measures to model non-stationary data to predict future values. The autocorrelation part of the model measures the dependency of a signal with a delayed copy of itself as a function. ARIMA models are capable of capturing temporal structures in time-series data. This model uses three parameters

LSTM-RNN[8] : Long Short Term Memory based Recurrent Neural Network. RNN is a generalization of feedforward neural network that has an internal memory. As its name suggests, it is recurrent in nature, i.e., it performs the same function for every input of data while the output of the current input depends on the past one computation. Once the output is generated, it is replicated and sent back to the RNN. Decisions are based on the current input and the previous step learnt output. LSTM is a modified version of RNN which makes it easier to remember past data in memory.

Transformer[9] : it is a neural network architecture that uses a self-attention mechanism, allowing the model to focus on the relevant parts of the time-series to improve prediction qualities. This consists of Single-head Attention and Multi-head Attention. The self-attention mechanism searches for inputs sequences, for each input, and adds result to the output sequence. These processes are parallelized allowing acceleration of the learning process.

b) Data

We have incorporated the daily OHLCV (Open, High, Low, Close, Volume) data of three companies Dr Reddys Laboratories Ltd, Wipro Limited, Reliance Steel & Aluminum Co for a time frame of 252 days, from 16Sept20 till 15Sept21. The data has been extracted & stored locally from NSE website. The data set incorporated is divided into a ratio of 7:3, where 70% of it will be the training data and the remaining for testing. For this project, we will ignore the top and bottom 5% of values, as they may be outliers.

c) Predicting Stocks

ARIMA: We initially scaled our features to avoid intensive computation. We imported the ARIMA module from statsmodels.tsa.arima.model to train our model. The ARIMA models uses the first 180 rows of the dataset as training data, with parameters p=2, d=1, q=0. These learnt results are utilized to predict the close price of the further rows of dataset (row number 180 onwards)

LSTM- RNN: We are developing a neural network regressor for continuous value prediction using LSTM. We initialize the same. Now, we add the first layer with the Dropout layer. The LSTM layer has 250 neurons that will capture the model trend. Since we need to add another layer to the model, our return_sequence is set to True.

The input_shape corresponds to the umber of time stamps and indicators. For the Dropout layer, a value of 0.2 denotes that 20% of the 250 neurons will be ignored. This is repeated for layers 2,3, and 4. Finally a one-dimension output layer is created. This layer is one dimensional as we are only predicting one price each time. We now compile our code using Adam optimizer and loss as mean square error. We then fit our model.

Fig 4: LSTM-RNN Implementation (File Name: rnn.py)

Transformer: To encode the notion of time into the model, time encoders have been incorporated in the code (class Time2Vector). The time encoder comprises of two main ideas periodic component (ReLU function) & non-periodic component (linear function). The class Time2Vector encompasses two functions build (initialize matrices) and call (calculates the periodic and linear time features). We now initialize the three Transformer Encoder layers Single Head Attention, Multi Head Attention, and Transformer Encoder layer.

(Note: Snippet of this is not in the report. Kindly refer attached python files to view them.) After these initializations, the training process begins, which is followed by testing the data. The parameters used are Batch Size: 12, Sequence Length: 3, and Size of model input: 5.

The stock_prediction.py imports all these models and returns the final outputs depending on the stock ticker symbol entered.

2) Alternative Approaches to Meanvariance Optimization (Maria Debora Braga) [10] This paper illustrates two examples risk-based asset allocation strategies: the global minimum variance strategy; and the optimal risk parity strategy.

Global Minimum-Variance Strategy the asset manager recommends a portfolio already known by the reader. It is the portfolio at the leftmost end of the efficient frontier, which, given the location, is the portfolio with the smallest attainable ex ante standard deviation. It is not necessary to implement Mean-Variance Optimization to identify the global minimum-variance portfolio (GMVP). It can be replaced with a simplified optimization algorithm with the formula for portfolio variance as an objective function to be minimized and the inclusion of the traditional long-only and budget constraints. Therefore, the optimization problem to be solved to identify the GMVP can be written formally as follows:

Optimal Risk Parity Strategy aims to overcome the problem of risk concentration. The alternative designation of optimal risk parity as equally weighted risk contribution strategy or portfolio suggests the criteria used to structure the portfolio coherently with the goal: to give each asset class a weight so that the amount of risk it contributes to overall portfolio risk is equal to the amount contributed by any other asset class in the portfolio. The above condition can become written formally as equality among component risks. Moreover, defined as the product of asset class weight in the portfolio and its marginal contribution to risk.

The identification of optimal asset class weights for a risk parity portfolio involves solving an optimization problem which is different from mean-variance optimization. More precisely, it includes a new objective function to be minimized together with the traditional constraints on portfolio weights.

portfolio theory that assumes that investors will make rational decisions about investments if they have complete information. A major assumption here is that investors seek low risk and high reward. The two main components of mean-variance analysis are as follows:

Variance Represents how spread out the numbers are in a set

Expected Return Probability expressing the estimated return of the investment in the security

If two different securities have the same expected return, but one has lower variance, the one with lower variance is the better pick. Similarly, if two different securities have approximately the same variance, the one with the higher return is the better pick.

Results & Conclusion

From our results of LSTM-RNN, and the Transformer when compared to the ARIMA benchmark, our conclusion is that the LSTM is better at getting the trends to almost replicate the original trend, but it does not react to directional changes as good as the Transformer which does not score so well when it comes to closeness to the original stock trend.

Any stock prediction model cannot provide greater accuracy without increasing the complexity of prediction operations exponentially. The key to optimizing a stock forecasting system is balancing accuracy and computational expense.

The rationale behind the use of ML is not just about novel architectures that revolutionizes forecasting and time-series prediction, but also the way each use case should be evaluated in terms of the domain of the use case itself. Referring to this specific use case, metrics are surely an evaluation factor, but it is the PNL that gives deeper insight into the accuracy of the model to predict changes based on the trends and patterns that are of significance and perform the best when looked at the big picture.

Identification of correct ML architecture in Active Portfolio Management is cardinal because the stock predicted expected returns will be the base for stock selection for portfolio optimization.

Also, from our research, we conclude that risk-based optimization techniques are superior to the traditional mean-variance technique. When the risk-based strategies become considered, the portfolio selection phase undergoes changes because the strategy comes up to a single risky portfolio proposal. Therefore, an investor does not have to choose an optimal point on the efficient frontier according to their risk tolerance/aversion and investment objective.

The decision-making process is therefore analogous to the Capital Asset Pricing Model. In CAPM, the same risky portfolio is offered to all the investors. A point is then selected on the Capital Market line. This is the line, in the risk-return space, that originates from the risk-free rate and is tangent to the efficient frontier. The tangency portfolio is also the portfolio with the maximum Sharpe ratio.

Written by Varun Chandra Gupta

References & Citations

[1] Portfolio Selection, Wiley Online Library [Online]

Available: https://onlinelibrary.wiley.com/doi/full/10.1111/j.1540-6261.1952.tb01525.x [Accessed Nov. 03, 2022]

[2] Portfolio Management, Investopedia

Available: https://www.investopedia.com/articles/07/portfolio-history.asp [Accessed Nov. 03, 2022]

[3] Portfolio Management, groww.in. [Online].

Available: https://groww.in/p/portfolio-management [Accessed Oct. 07, 2022]

[4] Derek Snow, Machine Learning in Asset ManagementPart 1: Portfolio ConstructionTrading Strategies in The Journal of Financial Data Science Winter 2020, [Online Document], Available: https://jfds.pm-research.com/content/2/1/10 [Accessed Nov. 03, 2022]

[5] Wei Chen, Haoyu Zhang, Mukesh Kumar Mehlawat and Lifen Jia, Meanvariance portfolio optimization using machine learning-based stock price prediction, ScienceDirect, [Online Document], Available:

https://reader.elsevier.com/reader/sd/pii/S1568494620308814?token=072C0F96416E333E05C4C88C 0F3EF956F66CB2B591F5865147EF13701FFC6CEB96D70EA6ADA889440EE8F47A7D90D1E6& originRegion=us-east-1&originCreation=20221007063440 [Accessed Oct. 07, 2022]

[6] ARIMA Model, ProjectPro [Online]

Available: https://www.projectpro.io/article/how-to-build-arima-model-inpython/544#:~:text=Model%20in%20Python%3F-,ARIMA%20Model%2D%20Complete%20Guide%20to%20Time%20Series%20Forecasting%20in% 20Python,data%20to%20predict%20future%20values [Accessed Nov. 03, 2022]

[7] ARIMA Model, TowardsDataScience [Online]

Available: https://towardsdatascience.com/time-series-forecasting-predicting-stock-prices-using-an arima-model-2e3b3080bd70 [Accessed Nov. 03, 2022]

Available: https://towardsdatascience.com/lstm-for-google-stock-price-prediction-e35f5cc84165 [Accessed Nov. 03, 2022]

[9] Stock Predictions with state-of-the-art Transformers and Time Embeddings, TowardsDataScience [Online]

Available: https://towardsdatascience.com/stock-predictions-with-state-of-the-art-transformer-and time-embeddings-3a4485237de6 [Accessed Nov. 03, 2022]

[10] Maria Debora Braga, Alternative Approaches to Traditional Mean-Variance Optimisation, SpringerLink, [Online Document], Available: https://link.springer.com/chapter/10.1007/978-3-319- 32796-9_6 [Accessed Nov. 03, 2022]

[11] Michael Pinelis, David Ruppert, Machine learning portfolio allocation, ScienceDirect, [Online Document], Available:

https://reader.elsevier.com/reader/sd/pii/S2405918821000155?token=E142E1AA4431A9B55ECF1C2 ABA5CC7E52CA69D5E1C728796D64FEC02BCFB0E3ABEEFAA56E0AF1E431BAE0CE7C041C E5E&originRegion=us-east-1&originCreation=20221007052506 [Accessed Oct. 07, 2022]

[12] Jrn Sass, Anna-KatharinaThs, Risk reduction and portfolio optimization using clustering methods, ScienceDirect, [Online Document], Available:https://reader.elsevier.com/reader/sd/pii/S2452306221001416?token=5E952856364B9133BB345373 B3373F4A2A5D62EAC2886F34758D28024D01C053253082DAC3408677BB6E425ECB264C6F& originRegion=us-east-1&originCreation=20221007063710 [Accessed Oct. 07, 2022]

How machine learning is used in portfolio management?

Follow this link:
How machine learning is used in portfolio management? - Rebellion Research

Read More..

Deep learning-based systems aid radiographers in prediction … – Omnia Health Insights

Artificial intelligence, and deep learning in particular, has been used extensively for image classification and segmentation, including on medical images for diagnosis and prognosis prediction. However, the use of deep learning in radiotherapy prognostic modelling is still limited.

Deep learning is a subset of machine learning and artificial intelligence that has a deep neural network with a structure like the human neural system and has been trained using big data. Deep learning narrows the gap between data acquisition and meaningful interpretation without explicit programming. It has so far outperformed most classification and regression methods and can automatically learn data representations for specific tasks.

The application areas of deep learning in radiation oncology include image segmentation and detection, image phenotyping and radiomic signature discovery, clinical outcome prediction, image dose quantification, dose-response modeling, radiation adaption, and image generation.

An article published in Clinical Oncology analyses 10 studies on the subject noting that researchers suffer from the same issues that plagued early normal tissue complication probability modelling, including small, single-institutional patient cohorts, lack of external validation, poor data and model reporting, use of late toxicity data without taking time-to-event into account, and nearly exclusive focus on clinician-reported complications.

It adds that the studies, however, demonstrate how radiation dose, imaging and clinical data may be technically integrated in convolutional neural networks-based models; and some studies explore how deep learning may help better understand spatial variation in radiosensitivity. In general, there are several issues specific to the intersection of radiotherapy outcome modelling and deep learning, for example, the translation of model developments into treatment plan optimisation that will require an additional combined effort from the radiation oncology and artificial intelligence communities.

Hence, the use of machine learning and other sophisticated models to aid in prediction and decision-making has become widely popular across a breadth of disciplines. Within the greater diagnostic radiology, radiation oncology, and medical physics communities promising work is being performed in tissue classification and cancer staging, outcome prediction, automated segmentation, treatment planning, and quality assurance as well as other areas.

Dr. Ali Vahedi, Consultant Radiologist at Mubadala Healthcare, explains that AI has an important role to play in the current and future of radiology. At present, this is primarily in the form of helping clinicians improve efficiency and diagnostic capacity which is essential with the exponential increase of diagnostic tests conducted year-on-year. He adds that AI has the potential to rapidly evaluate a vast quantity of imaging data, helping to prioritise worklists and diagnoses, which will aid in reducing reporting time, improve the accuracy of reports as well as limit discrepancies. In addition, it will give radiologists more time for direct patient care and vital research.

Dr. Ali Vahedi

A study published in the Progress in Medical Physics Journal highlights that high-quality simulated three-dimensional (3D) CT images are essential when creating radiation treatment plans because the electron density and anatomical information of tumours and OARs are required to calculate and optimise dose distributions.

Radiotherapy plays an increasingly dominant role in the comprehensive multidisciplinary management of cancer. As radiation therapy machines and treatment techniques become more advanced, the role of medical physicists that ensure patients safety becomes more prominent. With the advancement of deep learning, its powerful optimisation capability has shown remarkable applicability in various fields. Its utility in radiation oncology and other medical physics areas has been discussed and verified in several research papers. These research fields range from radiation therapy processes to QA, super-resolution medical image, material decomposition, and 2D dose distribution deconvolution.

According to Dr. Vahedi, the global imaging market size is expected to grow, driven by numerous factors such as growth in the number of hospitals and clinics and rising demand for minimally invasive surgeries.

He adds that rapid advancements in medical therapy also necessitate the need for regular multimodality imaging. In addition, technological advancements in medical imaging equipment are also contributing to the growth, with manufacturers introducing new products that are more compact, more cost-effective, and produce less ionising radiation than their predecessors. This improved affordability will invariably improve patient access to imaging. it can be concluded that over the past few years there has been a significant increase in both the interest in as well as the performance of deep learning techniques in this field.

Promising results have been obtained that demonstrate how deep learning-based systems can aid clinicians in their daily work, be it by reducing the time required, or the variability in segmentation, or by helping to predict treatment outcomes and toxicities. It remains to be seen when these techniques will be employed in routine clinical practice, but it seems warranted to assume that we will see AI contribute to improving radiotherapy soon. In conclusion, the application of deep learning has great potential in radiation oncology.

This article appears in the latest issue of Omnia Health Magazine.Read the full issue online today.

Back to Clinical

Excerpt from:
Deep learning-based systems aid radiographers in prediction ... - Omnia Health Insights

Read More..

Genomic Analysis, Machine Learning Leads to Chemotherapy … – GenomeWeb

A Hong Kong University of Science and Technology- and Samsung Medical Center-led team reporting in Genome Medicine outlines molecular features linked to chemotherapy resistance in forms of glioblastoma (GBM) containing wild-type isocitrate dehydrogenase (IDH-wt) genes. Using exome sequencing, GliomaSCAN targeted panel sequencing, and/or RNA sequencing combined with machine learning (ML), the researchers characterized molecular features found in patient-derived glioma stem-like cells (GSCs) originating from 29 temozolomide chemotherapy-resistant IDH-wild type GBM cases and 40 IDH-wild type GBM cases that were susceptible to the adjuvant chemotherapy treatment. In the process, they came up with a combined tumor signature coinciding with temozolomide chemotherapy response, highlighting a handful of genes with enhanced expression in the temozolomide-resistant cells, along with other chemotherapy resistance- or sensitivity-related features. With multisector temozolomide screening, meanwhile, the authors show that chemotherapy response can vary within tumor samples from the same individual. "We identified molecular characteristics associated with [temozolomide] sensitivity, and illustrate the potential clinical value of a ML model trained from pharmacogenomic profiling of patient-derived GSC against IDH-wt GBMs," the authors write.

View original post here:
Genomic Analysis, Machine Learning Leads to Chemotherapy ... - GenomeWeb

Read More..