

Feature Articles: Basic Research Envisioning Future Communication Vol. 14, No. 11, pp. 21–24, Nov. 2016. https://doi.org/10.53829/ntr201611fa3 Learning from a Large Number of Feature CombinationsAbstractSecondorder polynomial regression can often outperform simple linear regression by making use of feature combinations. However, when the number of feature combinations is large, secondorder polynomial regression quickly becomes impractical. In this article, we present convex factorization machines, a new technology developed by NTT Communication Science Laboratories, which can cope with a large number of feature combinations and guarantees globally optimal model parameters. Keywords: machine learning, regression analysis, feature combinations 1. IntroductionWith the democratization of the Internet, social media, and connected devices, the amount of data that can be used for scientific or business purposes is ever growing. In this context, machine learning has recently attracted a lot of attention due to its ability to leverage large amounts of data for predictive analytics. In particular, regression analysis is a frequently used predictive technology in machine learning. We present regression analysis by using house price prediction as a running example (Fig. 1). House price is typically determined by numerous features such as whether the house is detached or terraced (adjoined to other homes), the number of rooms, and whether it has a garden. We can use regression analysis to obtain from past examples of sold houses an equation that relates these features to the house price. In linear regression, the relationship between the features x = (x_{1},...,x_{d}) and the house price y is modeled by y = w_{j}x_{j} = w^{T}x, where w = (w_{1},...,w_{d}) is a weight vector estimated from previously sold houses. By inspecting the estimated weights, we can infer what features influence house price the most. In addition, by using the aforementioned model equation, we can predict the price of new houses, given their features.
However, while linear regression is very simple, it has some limitations. For example, while the price of both detached and terraced houses decreases with distance from the city center, we expect the price of terraced houses to decrease faster than that of detached houses. In this case, since linear regression estimates a weight for the distance from city center independently of whether a house is detached or terraced, it cannot achieve high predictive accuracy. To solve this problem, it is necessary to estimate different weights for the distance to the city center, depending on whether a house is detached or terraced. In other words, it is necessary to introduce feature combinations in the model equation. This is called secondorder polynomial regression. Secondorder polynomial regression can estimate models that fit the data better than linear regression. However, because the number of feature combinations is quadratic in the number of features, the number of feature combinations can quickly explode. For example, in genomic selection, which is the task of predicting grain yield from the DNA (deoxyribonucleic acid) of cereal plants, the number of genes is very large, and therefore, using feature combinations in the model equation can become impractical. Factorization machines (FM) [1] are a recently proposed method that can deal with a large number of feature combinations. Unfortunately, with FM, the quality of the estimated model strongly depends on the parameter initialization. To address this issue, we at NTT Communication Science Laboratories developed convex factorization machines (CFM), a new technology that can both cope with a large number of feature combinations and guarantees a globally optimal model regardless of the initialization [2]. 2. CFMIn secondorder polynomial regression, the relationship between house features and house price is modeled by the equation y = w^{T}x + x^{T}Wx, where again, w is a weight vector, and W is a matrix whose elements correspond to the weights of feature combinations. When the number of features d is large, estimating W can quickly become impractical because W is a d x d matrix. To address this issue, both CFM and the original FM reduce the number of parameters to be estimated by assuming that W is a lowrank matrix. With the original FM, W is replaced by PP^{T}, where P is a d x k matrix (k << d) and k is a userdefined rank hyperparameter. The original FM then use training data to estimate P instead of W (Fig. 2). However, because the estimation of P involves a nonconvex optimization problem, the quality of the obtained parameters greatly depends on the initialization. In practice, it is therefore necessary to try different initializations in order to obtain good results.
In contrast, our proposed technology, CFM, is guaranteed to obtain globally optimal model parameters regardless of the initialization. We developed an efficient algorithm to learn W in eigendecomposition form. We can use our algorithm to estimate the k eigenvalueeigenvector pairs of W (Fig. 3). In addition, our algorithm automatically determines the rank k of W from data.
In Table 1, we empirically compare ordinary secondorder polynomial regression (i.e., without a lowrank constraint), FM, and CFM on genomic selection (the task of predicting grain yield from the DNA of cereal plants). The values in the table indicate the Pearson correlation between the true grain yield and the grain yield predicted by the three methods (higher is better) on test data. Results for FM were obtained by trying several possible initializations. These results show that CFM can achieve higher predictive accuracy than FM. In addition, the CFM results are also better than those for ordinary secondorder polynomial regression. In machine learning, it is generally known that a model can overfit the data if the number of parameters is too large. We believe that CFM can mitigate this issue thanks to the reduced number of parameters to be estimated.
An important property of the lowrank constraint used in FM and CFM is that it enables the weights of feature combinations that were not observed in the training set to be estimated. This property is particularly useful in implementing recommender systems, a domain where FM have been particularly popular in recent years. 3. Higherorder extensionsWe presented CFM, a new technology capable of efficiently leveraging secondorder feature combinations. To further improve predictive accuracy, it is sometimes useful to consider thirdorder or higherorder feature combinations. We recently proposed new efficient algorithms for this purpose [3, 4]. References
