Innovativeness, Work Flexibility, and Place Characteristics: A Spatial Econometric and Machine Learning Approach

: This paper seeks to study work-related and geographical conditions under which innovativeness is stimulated through the analysis of individual and regional data dating from just prior to the smartphone age. As a result, by using the ISSP 2005 Work Orientations Survey, we are able to examine the role of work ﬂexibility, among other work-related conditions, in a relatively more traditional context that mostly excludes modern, smartphone-driven, remote-working practices. Our study conﬁrms that individual freedom in the work place, ﬂexible work hours, job security, living in suburban areas, low stress, private business activity, and the ability to take free time off work are important drivers of innovation. In particular, through a spatial econometric model, we identiﬁed an optimum level for weekly work time of about 36 h, which is supported by our ﬁndings from tree-based ensemble models. The originality of the present study is particularly due to its examination of innovative output rather than general productivity through the integration of person-level data on individual work conditions, in addition to its novel methodological approach which combines machine learning and spatial econometric ﬁndings.


Introduction
There is a great amount of strong empirical evidence for the existence of a relationship between innovation and economic growth. However, this relationship seems to depend on complex socioeconomic attributes, particularly in the context of regional economies [1,2]. Innovation itself, for its part, is the result of sophisticated interactions of social, economic, historical, and, not least of all, individual factors. Unequivocally, innovation is triggered and inspired via ideas developed by individuals, and the development of such ideas is subject to individual creativity.
A significant amount of empirical evidence indicates that the systems approach successfully explains the variations in innovation performance across locations and time periods (e.g., [3,4]). However, the focus on multiple actors and purposeful interactions tends to downplay the role of individuals and the "randomness" of innovative activities. Several studies have observed that innovative behaviour is induced by individuals' original ideas, and the road towards innovation performance and success is often shaped by individual creativity (e.g., [5][6][7]). Indeed, many patents are owned by individual inventors [8]. Furthermore, original ideas that result in innovative products or know-how are sometimes spontaneous and often originate from non-R&D-related activities outside firms [9][10][11][12].
In common, previous studies invariously adopt a human-resource approach and use firm-level data to investigate the work-innovation nexus. By contrast, this paper looks at work-innovation linkage from the perspective of individuals. On the basis of these observations, this study addresses the following questions: Which factors shape or determine creativeness? Is creativeness a product of the human mind determined by neural brain activity, or is the work and living environment of an individual a facilitating factor for creative ideas? In order to answer these questions, the present study will zoom in on the individual's work environment, as well as on regional factors that are likely to influence a person's creativity (see also [13]). In performing this study, we focus on the 2005-2006 period, which is just before the introduction of modern smartphones followed by tablets, which brought considerable remote and flexible working practices for many individuals [14,15]. While devices that can be defined as smartphones already existed by the period analyzed in this study, the arguably faster and efficient devices alongside with software such as the Apple iOS and the Google Android operating system had yet been introduced to global markets [14,16]. These technologies have enabled individuals to connect and engage in their work outside of the employer's locations at any time of the day, resulting in increased flexibility and organization of work hours [17,18]. As a result, the importance of location and the spatial dimension of inequalities in the accessibility of communication infrastructures with regard to regional industries, including the education sector, has become even more pronounced [19,20]. This newly induced type of flexibility may be involuntary or voluntary, even resulting in behaviors such as compulsive and continuous monitoring of work-related emails and blurring the borders between work and leisure times [16,17]. On the other hand, through our focus on a modern but pre-widespread smartphone use period, we are able to study the roles of job flexibility in a relatively more traditional sense without being subject to confounding elements arising from the prevalent usage of smartphones and tablets. In other words, job flexibility as measured in this study mostly captures "true" flexibility, rather than a type of flexibility caused by the voluntary or involuntary out-of-work habits. Consequently, using survey data from 18 European countries, this study examines the role of both individual work conditions and other placebased factors that influence regional innovation activities. The results from our analysis show that work-leisure balance and geographical attributes pertaining to city life and industrial clusters can significantly predict the innovation performance of a region. This result also holds after controlling for spatial spillovers and applying machine learning algorithms as an alternative methodological approach. In short, we find that flexible work time, job security, and job autonomy are conducive to regional innovation performances. Given our focus on regional innovation, this study provides additional evidence on the work-innovation linkage which hitherto has been mostly restricted to the firm level. As the RIS approach and the recent open innovation literature suggest, innovation often occurs in non-firm settings and involves actors outside the industry [21][22][23]. Our study addresses this gap and generalizes these insights at a regional level.
This study contributes to the area of innovation studies and the related human resources management HRM literature in three ways. Firstly, we adopt a workers' perspective in order to analyse the impact of work-related factors on innovation at the regional level. While there is a burgeoning literature on the impact of HRM practices on innovation performance (see Laursen et al. [24] for a survey of the literature), many studies adopt the firm's perspective and use firm-level data to understand the mechanisms underlying individual creativity and innovation performance. However, the use of this type of data limits the scope to only innovative activities that take place at the firm level. For example, regarding flexible working time, Arvanitis [25] finds that a higher level of individual worktime flexibility has a positive and statistically significant impact on innovation propensity and innovation success. Similarly, a study by Burtch et al. [26] suggests that the positive effects on innovation can be explained by the fact that HRM practices allow workers to redeploy resources strategically, thus stimulating innovative activities. Nevertheless, Arvanitis et al. [27] found that the impact of flexible work time is usually very small.
Secondly, this study integrates management innovation into the regional innovation systems RIS framework. Birkinshaw et al. [28] defined management innovation as a new management practice that furthers organisational goals. While this paper posits and finds that flexible work hours and a relaxing work environment are conducive to innovation, we also maintain that place-based factors underlying the RIS approach are also necessary to activate the HRM-innovation linkage. While workers may generate original ideas, the intellectual capital and knowledge base of a region, its university-industry linkage, its culture of idea sharing and cooperation, the rules that govern knowledge flow, and other factors that underlie the RIS approach are crucial in the process of turning a new idea into a patent or a new product. For instance, Kianto et al. [29] show that intellectual capital mediates the effect of knowledge-based HRM policies (e.g., concerning recruitment, training, and appraisal system) on innovation performance. As such, this study places the emerging literature on work-innovation linkage back in the context of RIS and enriches the RIS approach by adding a new layer to the systems tradition.
Finally, this study employs both spatial econometrics and machine learning (ML) techniques in the analysis. Machine learning techniques have increasingly been used in innovation studies. For example, Suominen et al. [30] used Latent Dirichlet Allocation (LDA) to show the evolution of the knowledge profiles of firms in the telecommunication industry. Ballestar et al. [31] applied Automated Nested Longitudinal Clustering (ANLC) and neural networks to evaluate the effect of a new employment policy on the productivity of university researchers in Madrid. Recently, Kim and Sohn [32] used semantic analysis to identify convergence patterns of new technology from patent data. However, to the best of our knowledge, this is the first study to complement findings from spatial econometrics with various machine learning analyses in order to study RIS and management innovation.
The remainder of the article is organized as follows. Section 2 outlines the influential contributions to the literature. Section 3 outlines the theoretical framework and provides the foundation for our econometric model. Next, Section 4 describes the data and visualizes the statistical distributions of key variables. Section 5 summarizes the findings from econometric estimations that serve as a reference for our machine learning techniques. These techniques are outlined in Section 6, which also presents several findings that are not covered by the preceding econometric analyses. Section 7 makes concluding remarks.

Literature Review
Ever since the seminal writings of Schumpeter [33], the theme of innovation has been high on the agenda of both the research community and decision makers in the public and private sectors. Numerous empirical studies have shown that innovation results in economic progress, while innovation itself is influenced by an array of factors and relations (see for instance [34,35]). The National Innovation Systems (NIS) and the RIS approaches describe innovation as an activity undertaken by a network of actors subject to an institutional environment [36,37]. The systems approach highlights the importance of interactions (e.g., learning, knowledge flow, and policies) among a network of actors such as universities, industries, and national or regional governments [38]. The differences between these interactive activities and institutions, which include, but are not limited to, financial capacity, organisational structure, social capital, and culture, determine the differences in innovation performance and economic growth across regions [39].
Although the topic of work-innovation nexus is relatively new in innovation studies, the role of individual creativity on innovation in professional work environments has been studied in the field of HRM. In an early study, Amabile et al. [40] argued that the social environment is likely to affect individual creativity and, therefore, influences the innovative output of an organization and its members. This social environment is characterized by five conceptual categories of work-related factors: an organizational culture that encourages risk-taking and creative attitudes; job autonomy that cultivates a sense of ownership and responsibility; the availability of sufficient resources; and stress from workload pressures and organizational impediments such as conservatism that limits and discourages personal creativity. More recent studies, however, have found only mixed evidence about the importance of these factors (see below).
A concept that is closely related to job flexibility is job autonomy. In this regard, Laursen et al. [41] and, more recently, Krammer [42] showed that employees' greater job autonomy contributes to the innovative performance of a firm (also see the literature review by Seeck and Diehl [43] on this relationship). Along similar lines, Arvanitis [25] found that greater decision-making power and span of control have positive impacts on both product and process innovation. Nevertheless, based on interviews with the executives of 49 technology small and medium-sized enterprises (SMEs) in Germany, Strobel et al. [44] reported that setting unclear tasks impedes innovative activity. As tasks related to innovation are not always well defined and require a high degree of cooperation, greater job autonomy could be harmful to innovation. In other words, the effects of task autonomy, flexible work hours, and job security on innovation remain unclear and more evidence is required. The present study contributes to the debate by providing additional confirmative evidence on the role of HRM practices.
Together with flexibility and autonomy, it has also been suggested that the level of security of an individual's job is also related to innovative performance. Michie et al. [45] found that labour market deregulation (e.g., increasing use of short-term or temporary contracts) has been detrimental to innovation in the UK. A more detailed examination by Zhou et al. [46] suggests that firms with high shares of workers on fixed-term contracts usually perform well regarding imitative new products but poorly in relation to innovative new products. More recently, Hoxa and Kleinknecht [47,48] has shown that greater job insecurity is associated with poorer innovation performance in German firms. On the contrary, evidence observed by Arvanitis [25] indicates that firms that offer low job security are likely to have more product innovations. Drawing on these views and findings, we formulate (in Section 3) a theoretical framework that is both based on the synthesis of previous research on innovation, creativity, and personal autonomy.

Theoretical Background
As has been stipulated several times in the literature, "innovation is not manna from heaven" [49], but it is a product of knowledge-based human activity. In other words, we may hypothesize a relationship that explains how cognitive-oriented innovative performance is determined by various background factors (such as work stress, work effort, job cooperation, etc.). Celbis and Turkeli had shown that at the country level, innovation exhibits diminishing returns on work hours [50]. A similar relationship between leisure time and productivity in general is observed by Cui et al. [51]. We also assume that innovative performance does not increase linearly with respect to the number of work hours, but that there is some sort of inverted U-shape relationship. Consequently, we would have to look for the optimal work effort that maximizes innovativeness. We introduce this work-time related factor in the econometric specification of a knowledge production function, based on the seminal work by Jones [52]. We estimate the model in Section 5 alongside other variables that represent the degree of work-hours flexibility, freedom to organize one's own work, ability to take time off, job security, and other personal and regional attributes that are explained later in this section.
We assume that all individuals, regardless of industry class and education level, can be creative idea-seekers if they are given the opportunity. Nevertheless, the generation of new knowledge by idea seekers in a region also depends on the existing stock of knowledge, which is not only specific to that region but exists in the country (or in the world) as a whole. We augment knowledge production function in two ways. Firstly, we introduce a set of individual level work-related factors and relevant regional characteristics. Secondly, as we assume that innovativeness may exhibit diminishing returns on work hours, we introduce a quadratic specification, followed by an alternative specification that considers a "deviation from the ideal work hours" effect. In other words, we assume that there exists an ideal length of work time that allows the idea seekers to work at their full potential, thus maximising regional innovativeness on average. When a region deviates from this optimum length of work time (either positively or negatively), its innovativeness will decay, and this decay increases exponentially.
Of course, the length of work time alone is not sufficient for understanding the conditions that shape a work-life balance that may be related to intellectual output. In order to account for time outside of work hours, apart from simply the length of work time, we also include, in our analysis, variables representing job flexibility (Flexible), freedom to organise daily work (Free), and ability to take time off (TimeOff).
The amount of stress to which individuals are subject may also have an impact on their creativity [53]. Furthermore, individuals may face stress resulting from low income, unemployment, and lack of job security. Therefore, it is plausible to include also additional explanatory variables (NeverStress, SomeStress, LowIncome, Unemployed, and NotSecure) that represent these attributes, as defined in Table 1.
In addition to factors related to work hours, some location-specific characteristics of the areas (regions and urban areas) where individuals reside may also be potential determinants of personal innovativeness. In a recent OECD report, it was shown that productivity is particularly driven by second-tier cities compared to core cities [54]. Furthermore, regional productivity in general is subject to local market access and agglomeration levels [55,56]. The International Social Survey Programme (ISSP) Work Orientations Survey (ISSP, 2005) provides classifications of the regions, i.e., areas, in which the respondents reside, and we accordingly distinguished the types of urban areas in our analysis, viz., large urban areas (BigUrb) and suburban areas (SubUrb) versus the reference category that consists of smaller settlements.
Past research has shown that entrepreneurship and human capital are key determinants of innovative activity and crucial factors of development in lagging areas [57][58][59][60]. Together with human capital, measured by the level of education (Degree), we include variables that measure the importance of entrepreneurs (Self) and of private industry (Private) within the model. In conjunction with industry characteristics, research and development (R&D) activities are proxied by the share of individuals in science-related professions (Science) and the health industry (Health).
We note here that the nexus of knowledge-based creativity and the determinants of individual and contextual factors may call for complementary model adjustments, depending on data availability. It may, therefore, be necessary to pursue further empirical alterations to the conceptual model sketched in this section before it can be used for estimation in our econometric models and for prediction in our machine learning models. Furthermore, innovative activities are not constrained by country or regional borders and usually involve international collaboration. In order to account for knowledge spillovers across borders, a consideration of spatial dependence will, thus, be pertinent. These issues are addressed in more detail in Section 5. Share of respondents who are either completely free to decide the start and end times of their work or who have some freedom to decide within certain limits.
Share of respondents whose starting and finishing times are decided by the employer. Source: ISSP 2005-Work Orientations III.

Free
Share of respondents who have complete or some freedom to organise their daily work.
Share of respondents who have no freedom to organise their daily work. Source: ISSP 2005-Work Orientations III.

NotSecure
Share of respondents who either disagree or strongly disagree that their job is secure.
Share of respondents who either agree, strongly agree, or neither agree or disagree that their job is secure. Source: ISSP 2005-Work Orientations III.

NeverStress
Share of respondents who state that their work is "never" or "hardly ever" stressful.
Share of respondents who state that their work is "always" or "often" stressful (same reference dummy as the variable SomeStress). Source: ISSP 2005-Work Orientations III.

SomeStress
Share of respondents who state that their work is "sometimes" stressful.
Share of respondents who state that their work is "always" or "often" stressful (same reference dummy as the variable NeverStress). Source: ISSP 2005-Work Orientations III.

Science
The share of respondents whose occupations belong to the category "physical, mathematical and engineering science professionals" listed under the International Share of respondents who either disagree or strongly disagree that their income is high.
Share of respondents who either agree, strongly agree, or neither agree or disagree that their income is high. Source: ISSP 2005-Work Orientations III.

BigUrb
Share of respondents who live in an urbanised area/big city.
Share of respondents who live in either a town or small city, country village, or farm or home in the country (same reference dummy as the variable SuBurb). Source: ISSP 2005-Work Orientations III.

SubUrb
Share of respondents who live in a suburban area or on the outskirts of a big city.
Share of respondents who live in either a town or small city, country village, or farm or home in the country (same reference dummy as the variable BigUrb

Data and Descriptive Statistics
Since, to the best of our knowledge, there are no studies that systematically investigate the relationship between individual work conditions and innovative activities at the regional level, there is no database that contains sufficient information on both variables. For our analysis, this limitation has necessitated the merging and aggregation of data from different sources. We obtained the data for the dependent variable from Eurostat, while all remaining data on the explanatory variables are taken from the International Social Survey Programme [61].
Challenges related to the compatibility of heterogeneous databases have influenced a number of our decisions regarding the design of research, as detailed later in this section, particularly those concerning the choice of the units of analysis and the measure of innovativeness.
The Work Orientations module III of the ISSP provides a number of interesting variables that allow us to test our hypotheses. The main advantage is that the survey asked about individuals' attitudes towards work, work conditions, and employment arrangements. Table 1 presents the set of variables that we obtained from the ISSP survey and their descriptions, together with the reference categories that apply to certain variables. The descriptive statistics for each of these variables are displayed in Table 2.
The ISSP 2005 survey included 32 countries, 19 of them being in Europe. Unfortunately, as the regional classification in the ISSP is not always consistent with the Nomenclature of Territorial Units for Statistics (NUTS) classification of Eurostat, we have not been able to merge data for Norway. Therefore, the analysis comprises 18 European countries: Belgium, Bulgaria, Cyprus, Czechia, Denmark, Finland, France, Germany, Hungary, Ireland, Latvia, the Netherlands, Portugal, Slovenia, Spain, Sweden, Switzerland, and the United Kingdom. Due to similar compatibility issues, we had to regroup some of the regions based on the NUTS classification.
Measuring innovative activities is another major empirical challenge, and there is a lack of consensus in the literature on how this can be achieved. In our study, we have considered R&D expenditures, but the data points are missing for many of the regions. The Regional Innovation Scoreboard (RIS) provides an alternative. However, the data for the year 2006 are not available in the RIS. As a result, following Buesa et al. [62] and Bergamini et al. [63], we used patent applications as a proxy for regional innovativeness. The data for patent applications originate from Eurostat and are included in our analysis in per capita terms so as to adjust for scale effects. As a result, the dependent variable used in the present study (patents per capita) has 264 observations. When the 16 explanatory variables are included, the dataset attains 253 complete rows. The definition of all variables and their summary statistics are included, respectively, in Tables 1 and 2. In principle, one could use the number of patents that an individual applies for as a proxy of innovation, but owing to privacy issues, in general, it is not possible to obtain information related to the inventors' time use and work flexibility. Therefore, prior to constructing econometric and machine learning models, we had to use regions instead of individuals as the units of analysis. This required us to aggregate (or average) our indicators for work conditions and other individual-level variables in order to match them with the patent's variable. More specifically, as pointed out in Section 3, most individual responses from the ISSP are expressed as proportions, and this affects the interpretation of the effects of the explanatory variables.
A visual representation of the spatial structure of our data is presented in Figure 1 where the circles mark the regions in the sample, and larger circles indicate larger sample sizes. A country-level categorisation of regions is provided in Figure 2: Switzerland (CH) and Sweden (SE) had the highest level of per capita patents in 2005, greatly diverging from the rest of the countries in our sample. Values shown in light green indicate less than average work hours, and the numerical value is also reported on top of each bar. The values in red are higher than the optimum work hours that we estimate in Section 5 (about 36 h per week), while the values coloured in darker green are those that are below this level. For instance, the Netherlands (NL) has the lowest average length of work time, whereas Bulgaria (BG) and Czechia (CZ) have the highest volume of work hours but very low levels of innovativeness.    Figure 2. Patents per capita and hours by country. Figure 3 is a plot of country-level per capita patents and work hours (on a standardised scale). The figure shows that Switzerland and Sweden are relatively close to the optimum level of work hours. However, there are other countries, such as Cyprus (CY), with lower innovativeness that are also around the same level. This is not unexpected, as it is clear that work hours alone do not explain differences in innovativeness, an issue that we examine in further detail in Sections 5 and 6. As shown earlier in Figure 2, Bulgaria and Czechia are countries with high working hours but low innovativeness. Figure 4 displays the same relationship at the regional level. For the sake of readability, we label only those regions that are more than 1.2 standard deviations away from the average hours and those that are more than three standard deviations higher than average per capita patents. The top part of the scatterplot is dotted with highly innovative Swiss regions that are mostly close either to the average hours or to the estimated ideal hours. The Dutch region of Noord-Brabant stands out as a highly innovative region with low working hours (the region of Noord-Brabant has been an important innovation hot spot in the Netherlands [35]). Figure 4 also shows that there is a considerable number of regions with either too long or too short working hours along with low innovativeness.
Finally, the levels of innovativeness of regions-ordered by country from high to low-is shown in Figure 5 where darker region labels indicate longer working hours. For each country, only the regions that are in the first, second, and third quartiles and those with the minimum and maximum per capita patent values are shown, provided that they exist in our dataset. As shown earlier in Figure 3, the Netherlands has regions where work hours are low but innovativeness is high. The darker region labels (those with a very high work time effort) are mainly positioned on the right side of the graph where per capita patent values are lower.  In general, the figures presented here hint at a non-linear relationship between work time and innovativeness. This being said, there are clearly other attributes related to the use of work time, in addition to simply the duration of work, e.g., flexibility, freedom to organise one's workload, and ability to take time off among many other work and locationrelated factors. Consequently, it is likely that there is a wealth of moderator variables that may impact on creativeness. However, the number of these explanatory variables is simply too large to be meaningfully treated in a standard econometric model; therefore, we will use a simple econometric model as a stepping stone for a more advanced machine learning approach. We aim to revisit and examine the roles of such a larger set of variables through machine learning models in Section 6 by considering the econometric output from Section 5 as a reference point.

Spatial Econometric Estimation Results
In this section, we present the results of the econometric model that corresponds to the conceptual model discussed in Section 3. Given the complexity and multidimensionality of the underlying causality patterns, we apply this model later in a complementary machine learning approach, after the estimation of our first-stage regression procedure. As a meaningful point of reference for the statistical performance of the machine learning models, we first estimate a linear regression model augmented with a quadratic term and spatial effects. The quadratic term reflects the inverse U-shaped function discussed in Section 3. Furthermore, this first, non-spatial specification assumes that the observations in one region are independent of observations made in other regions such that if is the error term of the model, then the assumption is E( i j ) = E( i )E( j ) = 0 for two regions i and j [64]. On the other hand, when spatial interdependencies among i and j are considered, for instance, in the case of the outcomes y i and y j , the latter would enter as an explanatory variable for the equation that pertains to the former and vice versa [64]. Considering all locations in the data, the spatial dependencies across all regions can be represented in a spatial autoregressive process by using a spatial weight matrix, as will be detailed in Equation (1) [65]. A spatial model assumes that the inclusion of this spatial autoregressive term ensures that the E(ε i ε j ) is equal to zero where ε is the error term of the spatial model and that ε ∼ N(0, σ 2 ) [64,65].
We begin by implementing a spatial model that is naive in the sense that it ignores endogeneity concerns and accounts only for spatial effects in a simple-and probably misrepresented-form. The former of these shortcomings is basically caused by the fact that all work-related variables from the ISSP survey are already included, and the prospect of successfully finding suitable instrumental variables for them is uncertain, particularly in a cross-sectional setting. The latter drawback is due to the fact that there are some administrative areas among the regions in our sample that are not included in our data. These missing observations would then bias our estimates of indirect spatial effects, and estimating more sophisticated model specifications such as a Spatial Durbin Model in such a context would still be inadequate. Rather than working with econometric models that are semantically not well specified, we will address these concerns in the machine learning (ML) models where they are internalised with a black-box approach. As for simultaneity concerns-i.e., a causal effect of per capita patents on the explanatory variables-we highlight here the fact that there is no clear automatic mechanism that would pressure the managers in a region, on average and from all sectors, to reduce work hours or to increase job flexibility due to a higher number of patents per capita in that region. In essence, this study seeks to acquire new insight into two main elements of information from the econo-metric estimations for referencing and cross-checking in machine learning procedures: the signs of the coefficient estimates and the optimum number of work hours.
The aforementioned spatial effects are particularly relevant within the context of innovation, growth, and development [66][67][68]. This relevance was theoretically formalised by Ertur and Koch [69], who showed that spatial externalities resulting from the local accumulation of knowledge result in interdependencies among countries. Knowledge spillovers have been observed with greater precision when examined on a disaggregated spatial scale, particularly at the regional level (e.g., [70][71][72][73][74], among others.) Petruzzelli [75] has shown that knowledge spillovers depend on the relative locations of firms and universities in a region. In order to account for spatial dependence in innovativeness, we specify a Spatial Autoregressive Model (SAR): where y is an N × 1 vector of the natural logarithms of per capita patent applications, and N is the number of observations (N = 253 regions); it is an N × K matrix consisting of K explanatory variables, together with country dummies. The set of variables in X includes explanatory variables that may or may not have a linear effect. However, a "deviation from the optimum" type of specification for these variables (that is, assuming the existence of an optimum level) would not be as straightforward as it is for the case of working hours. The reason for this is that these variables are obtained from individual-level survey responses and subsequently aggregated-or averaged if necessary-to a regional level. Therefore, most of these variables are not individual-level continuous quantitative indicators. Instead, they measure the region-specific share of respondents who responded in a certain manner to a particular question. For instance, as shown in Table 1, the variable Flexible is defined as "the share of respondents who are either completely free to decide the start and end times of their work, or those who have some freedom to decide within certain limits". The labels, definitions, and the reference categories (when applicable) of all variables are presented in Table 1. Finally, the country-level stock of knowledge, together with other country-level factors such as institutional quality and respect for the rule of law, is represented in the model by the inclusion of country dummies in all subsequent econometric estimations and data-analytical exercises. µı n is a constant term, and σ 2 is a constant variance parameter. The work hour variable is introduced in a quadratic form such that its effect is represented in two components: the coefficients of Hours and Hours 2 .
In Equation (1), the degree of spatial spillovers of regional innovativeness is evaluated by the parameter estimate ρ. The spatial connectivity is modelled through W, which is an N × N spatial weight matrix. Each element w ij of W is the inverse of the Euclidean distance from the main urban centre of region i to that of region j, and w ij = 0 if i = j where i = 1, . . . , N and j = 1, . . . , N. The aforementioned quadratic definition of the work hours variable aims to capture the potential non-linear effect of this variable and to estimate the ideal length of work time, as outlined in Section 3. The effect of deviating from this optimum length is estimated after replacing the quadratic term with the absolute value of the mean regional deviation from the optimum number of work hours denoted by |z|. The estimation of a SAR model using ordinary least squares (OLS) would result in biased and inconsistent estimators [76]. To cope with this problem, a maximum likelihood estimation (MLE) that enables asymptotic efficiency is recommended for addressing the bias and inconsistency issues in spatial models [76,77]. That being said, reporting SAR results should be conducted with caution: the direct effect of an explanatory variable appertaining to a certain region on that region's own innovativeness should be distinguished from the indirect effect of the same variable on the levels of innovativeness in other regions. To be more specific, we follow Elhorst [78] who shows how the direct effects for Equation (1) are calculated as the average of the diagonal elements of (I − ρW) −1 β k , which is the partial derivatives matrix with respect to the kth explanatory variable. The indirect effects of k, on the other hand, are given by the mean row (or column) sum of the off-diagonal elements of the same matrix [78]. The direct effects are the same as the SAR coefficients reported in Table 3 up to the second decimal digit, and since there are no significant indirect effects of the explanatory variables, we will simply report the SAR coefficients(the direct, indirect, and total effect breakdown can be provided upon requested). Finally, we also estimate a non-spatial version of Equation (1) (assuming ρ = 0) using OLS, and we report the results in the first two columns of Table 3.

Results of the Spatial and Non-Spatial Models
The third and fourth columns of Table 3 present the estimation results of the full model as shown in Equation (1). For both specifications, the results with the quadratic work hour variable are presented first, followed by the estimation results with the deviation from the optimum hours instead. For generating the |z| variable, we use the optimum work hours as estimated by the SAR specification, which is 35.7 h per week.
Firstly, the estimation results in Table 3 highlight a non-linear effect of average regional work hours on innovativeness; work hours affect regional innovativeness positively up to a certain level, but this effect is reversed if work time is too long. The SAR and the non-spatial results are very similar, with the estimate of the optimum work time length being 35.7 h according to the former. The non-linear impact can also be observed when the effect of work hours is expressed as a deviation from this optimum length: Regional innovativeness is reduced as average work hours exceed or fall behind the ideal level (the coefficients estimates on |z|). The estimated coefficient suggests that a regional average deviation of one hour reduces the per capita patent applications by 5%.
In contrast to what is found by Beugelsdijk [79], several other variables related to work time such as Flexible, Free, and TimeOff do not appear to yield significant coefficient estimates (however, see [27] for a different perspective). We note here, however, that the machine learning approaches in Section 6 may provide more information regarding the importance of these variables. Nevertheless, the estimation results provide clearer evidence regarding the following: the positive role of the prevalence of the private sector [80]; the share of highly educated individuals [29,81]; the level of urbanization [82,83]; and the weight of the scientific professions in the labor market [84] of a region. Spatial dependence in innovativeness-which should not be over-interpreted, as discussed earlier in this section-is significant and positive, suggesting that there are positive spillovers of knowledge across regions. The associated likelihood ratio and Wald test p-values highlight the significance of the spatial processes in innovativeness and indicate that the spatial specifications should not be reduced to a non-spatial model.
Considering the spatial econometric results presented in Table 3 as a frame of reference, in the next section we aim to provide a deeper understanding of the complex and multidimensional mechanisms that shape regional innovativeness through the application of a collection of machine learning techniques. Notes: Robust standard errors in parentheses: * p < 0.10, ** p < 0.05, *** p < 0.01. To preserve the validity of the reference categories, for each survey variable, the corresponding share of respondents who did not answer is included, but their parameter estimates are not reported.

Machine Learning Applications
Moving from a conventional model-based econometric approach to an algorithmic modelling framework is a big step, as heuristics-rather than an explanatory conceptualisation-comes into play. In fact, the apparent disconnection between the two approaches in the literature was famously outlined by Breiman [85] as the "two cultures" of modeling as opposed to algorithmic learning. While the different cultures are still apparent in economics, over the years the disconnection has lessened, and machine learning has become more accepted in the analysis of economic data [86]. There is a wealth of machine learning (ML) techniques in the current AI literature. Here, we apply five tree-based supervised machine learning methods on our data. Unlike the econometric estimations presented above, ML models necessitate an initial partitioning of the data into two (further discussed below). In order to preserve comparability between the econometric and machine learning applications, we replicate the earlier econometric estimations using the same (partitioned) data as in the machine learning models and compare their root mean squared errors (RMSEs). To further ensure comparability, we run two versions of each machine learning procedure. Firstly, we omit country and spatial effects from the data sets in order to obtain a clearer understanding of strong predictors. Secondly, we rerun the machine learning models including the country dummies and the longitude and latitude values that were used earlier to create the spatial weight matrix W.

The Base Regression Tree Model
The model of a single regression tree, sketched and applied in this section, serves as the central component of the more powerful ensemble methods that we implement in Sections 6.2-6.4 (the following R packages were used for applying and visualizing the ML models: rpart by Atkinson and Therneau [87]; randomForest by Breiman and Cutler [88]; xgboost by Chen et al. [89]; rpart.plot by Milborrow [90]; and pdp by Greenwell [91]). A complex, non-linear relationship between the explanatory variables and the dependent variable can be explained better by a regression tree approach compared with a linear regression model, particularly if the causal relationship between these factors is not well approximated by the latter [92,93]. Classification and Regression Trees (CART) provide predictions by splitting the data into subsets [94]. Since the dependent variable is continuous, our models are based on regression trees instead of classification trees. At first, our data are randomly split into two parts: the training data and the test data with shares of 70% and 30%, respectively. Leaving out the test data, a single regression tree uses "recursive binary splitting" in dividing the training data successively in a top-down manner, beginning with an initial split of the data into two regions R 1 (k, s) = {X|X k < s} and R 2 (k, s) = {X|X k ≥ s}, where X k is a selected predictor (out of the earlier-defined K explanatory variables) from the feature space X = (X 1 , . . . , X K ), and s is its value used as the first split point [95]. The purpose is then to identify the predictor and splitting value that yields the minimum total of the sum of squared residuals of regions R 1 and R 2 [92,95]: where y i represents the patent applications per capita for region i; andȳ R 1 andȳ R 2 are, respectively, the mean values of the same variable for the two data regions (the index term R and the word "region" in this context refers to specific parts or portions of the data split by the ML algorithms and should not be confused with the actual units of observations which are European regions). As the minimisation of the sum of squared residuals (SSR) is performed with respect to the predicted value rather than a slope coefficient, in this framework, the SSR is simply the squared deviations from the mean (see Equation (9).11 in [95]). In other words, the split point decisions aim to minimise the total sum of squares (TSS) (in a tree-based machine learning context, the dependent variable is often called the "outcome", while the set of explanatory variables, also referred as predictors, make up the "feature space"). Following the initial partitioning of the data, the next step of the recursive binary splitting process is the application of an iterated version of the minimisation in expression (2) to each of the newly identified regions R 1 and R 2 . As a result, the total number of terminal nodes (leaves) R j in this step becomes four. This recursive splitting process stops when a terminal node has too few observations but continues for those regions with a sufficient amount of data. This process yields a large tree that needs to be pruned in order to prevent the overfitting of the data [92,95]. A tree with the highest possible number of splits (i.e., with the highest degree of complexity) would be a full description of the data concerned, resulting in overfitting in a similar manner to a regression where the number of variables is equal to the number of observations, which yields poor out-of-sample predictions [93]. In any case, splitting the data in such a way-which is analogous to using all possible variable interactions in the context of a global parametric equation such as a traditional frequentist regression model-would clearly be a troublesome task to handle the assumption [96][97][98]. A regression tree model, on the other hand, searches automatically for the best interactions and can, therefore, be observed as a "highly interactive function class" due to its local and nonparametric structure [99,100]. The aforementioned pruning process necessary to prevent the overfitting of such a highly interactive specification involves the penalisation (also called "regularisation") of the size of a tree, measured by its number of terminal nodes |M|. Upon incorporating the complexity parameter (τ) into the model, the fully grown tree is pruned by the minimisation of the following: where m is the index of the terminal nodes (m = 1, . . . , |M|); and R m is the group of observations that fall into the m'th leaf [92,95]. Therefore, since the error of the treesummarised by the first term in expression (3)-is inversely related to |M|, the purpose is to find a balance such that the error is as low as possible without fitting an excessively complex tree [92,95]. An optimal value for τ can now be selected following a "K-fold cross-validation" process-as outlined in [92,95]-which randomly splits the observations in the training data into l equal sections (l = 1, . . . , L) (standard notation indexes the "K" folds conveniently by k. We use notation l in order to distinguish this index from the index for the explanatory variables (the K predictors)). Prior to this step, the corresponding unique subtrees that minimise expression (3) are identified for an array of possible values for τ. This step is also applied l times using all the training data except the lth section, and the mean squared prediction errors (MSEs) on the left-out observations are calculated for each level of τ. Due to the fact that in-sample assessment can overstate model accuracy, excluding portions of the data to evaluate model performance is particularly essential [99]. The L number of MSEs are averaged for each level of τ as CV τ = 1 L ∑ L l=1 MSE (l,τ) , where CV τ is the cross-validation error function. Finally, the unique subtree of the initially generated unpruned large tree (which uses the full training data) corresponding to the complexity parameter τ that minimises the cross-validation error is used to predict test data [92,95].
The straightforward visual representation of a regression tree allows for a relatively effortless interpretation. However, this method typically has lower predictive accuracy than regression models and may not be robust to changes in the data (i.e., it has a high variance) [92]. Furthermore, in the case of high correlation between model covariates, explanatory variables with a strong relationship to the dependent variable may end up being omitted by the tree model because the splits can be made using other inputs that are highly correlated to those strong predictors [101].
For a preliminary look at the prominence of the variables and the structure of their relationship to per capita patents, we first produce a single regression tree. In addition to the complexity parameter, tuning is required for the minimum number of observations in a terminal node and the maximum depth of the tree (i.e., the maximum number of nodes between the leaves and the root). We conduct a grid search for jointly determining the optimal values for these parameters. The minimum 10-fold cross-validation error corresponds to τ = 0.034, with a minimum number of five observations and a maximum depth of nine. All tree-based models that we apply use the deviation version of the work duration measure (|z|), exclude the quadratic expression for Hours, and include all categories for each predictor (i.e., the exclusion of reference categories is not required).

Regression Tree Results
The single regression tree is presented in Figure 6 where an initial split is made based on Flexible. The predicted level of innovativeness is the least in regions where less than 26% of individuals do not have any work-hours flexibility. Two other work-time related predictors, Free and |z|, also appear as splitting features. In accordance with our econometric results, innovativeness is predicted to be lower in regions where average work time diverges by about 3 h or more from the optimum level. Additionally, the tree predicts higher innovativeness in regions with a higher share of individuals who have some freedom in their daily work organisation. The model results also reinforce the validity of the econometric findings regarding the positive role of private businesses in reinforcing innovativeness. Finally and not surprisingly, this basic single tree approach suggests that low income is associated with lower regional innovation levels. When the regression tree is constructed after adding country dummies and the longitude and latitude variables in order to increase comparability with the econometric results, it yields an RMSE value of 174.16. When the econometric estimations are replicated using the same training data and when their prediction performances are tested on the test data, we observed RMSEs ranging from about 110 to 115.  In order to alleviate the above-mentioned drawbacks of using just one regression tree, we now proceed by applying certain appropriate ensemble and sequential techniques for the prediction of per capita patent applications, as outlined in Sections 6.2-6.4. Ensemble methods combine the predictions of many regression trees resulting in greatly improved predictions [92]. The ensemble methods that we apply on our data are bootstrap aggregation, random forest, and gradient boosting machine learning techniques.

Bootstrap Aggregation
If the regression tree method is applied on different data sets, such as a new training data set resulting from an alternative random partitioning of the same original data, trees with very different structures and predictions may be reached. This is referred to as "the model having high variance" [92]. The bootstrap aggregation method [102], also called "bagging," averages over multiple versions of the predicted outcome, each based on a different bootstrap sample drawn with replacement. In our case, bootstrap aggregation is performed by drawing B samples (b = 1, . . . , B) of size N, which is the total number of European regions in the training data. Subsequently, the mean predicted per capita patent application value is calculated. By averaging the outcomes of many tree models-each corresponding to a separate sample b of size N-it is possible to construct a tree-based statistical learning model with a low variance [92]. The bagging estimate of patents per capita for region i is given byŷ is the predicted per capita patents for region i resulting from the unpruned unique tree produced by the bth bootstrap sample [92,95].
For each region i in our training data set, there are a number of bootstrap samples (around B 3 ) in which that region is not included [92,103]. In other words, every region is an out of bag (OOB) observation in about one-third of the bootstrap samples [92]. A single estimate of patents per capita for every region can be obtained by averaging their predictions resulting from the samples in which i is OOB (called OOB predictions). In line with this, Svetnik [104] used these OOB predictions to estimate the error rate of the full ensemble.
Ensemble tree methods combine the degree of prominence of each variable resulting from each unique tree by summarising their contribution to the accuracy of the prediction in the form that is called a "variable importance" measure. In a bootstrap aggregation model, the degree of importance of a feature X k is calculated as the mean of the tree-specific decreases in the total of the sum of squared residuals resulting from all binary splits over X k . This degree of importance is then scaled into a 1-100 interval where 100 corresponds to the variable with the highest importance [92]. The outcome of a search for the number of trees that yields the minimum estimated test error (OOB-MSE) suggests using 173 bagged regression trees for obtaining y bag .

Bootstrap Aggregation Results
The resulting importance levels of the variables for the bootstrap aggregation model are presented in Figure 7. The flexibility of work hours and the prevalence of the private sector show up as the most important variables. Being in suburban areas and having a low income also appear to be influential in explaining innovativeness. In addition to Flexible, other variables related to work conditions such as Free, TimeOff, and NeverStress strongly contribute to the prediction of per capita patents, along with the other features with lower levels of importance.
When the country and spatial variables are reintroduced, the optimum number of trees becomes 10, and the model yields an RMSE of 137.2214, which is a clear improvement over the single tree model albeit with slightly worse predictions than the econometric results.

Random Forest
It is reasonable to expect a high degree of correlation among the trees built by the bootstrap aggregation technique, as variables with strong explanatory power will almost always appear as splitters in the top nodes of every tree [92,95]. If the trees are positively correlated, the variance of the average predicted per capita patent values will be ς 2 + 1− B ς 2 (instead of only ς 2 B , where ς 2 is the variance for an outcome estimate that is i.i.d.), which can be reduced either by increasing the number of bootstrap sample trees (a higher B) or by reducing correlation coefficient . The random forest approach, developed by Breiman [103], augments bootstrap aggregation by aiming for the latter before averaging the predictions from each tree for each observation [95,103].
As in the bootstrap aggregation approach, a random forest procedure makes use of a large ensemble of unpruned trees, each based on a bootstrap sample b drawn from the training set. On top of the randomisation introduced by bagging, a random forest procedure introduces further stochasticity by considering only a randomly chosen subset from the feature space at each node of the recursive binary splitting process. The rule of thumb is to calculate-in each binary split step-the best predictor-splitting value combination using only a random sample of q explanatory variables out of the full set of K features, where q = K 3 with a minimum node size of five [95,103]. As in bagging, a random forest procedure evaluates variable importance. Each observation not selected by sample b (i.e., each OOB observation of b) is fitted into the tree corresponding to this sample, and the accuracy of the prediction and the average decrease in MSE is noted. Subsequently, one at a time, the values of each predictor are randomisedkeeping the values of the remaining features untouched-and then the decrease in precision as a result of this randomisation is calculated. The loss in accuracy corresponding to a predictor X k is averaged over all trees B, yielding the importance value of its random forest-based variable [103]. Our random forest model, composed of 500 trees with K = 18 and q = 6 and minimum node size of 5, provides further useful information.

Random Forest Results
As presented in the variable importance plot in Figure 8, the random forest results strongly highlight the necessity of flexible work hours. In conjunction with the econometric findings, which provide knowledge regarding whether the predictors are positively or negatively associated with per capita patents, we observe further evidence on the contribution of living in suburban areas, income, the private sector, and the ability to take time off. This random forest model, when augmented with the country and spatial variables, performs with an RMSE of 131. 85.
The random forest model can also be used to calculate an inter-observation proximity measure. Returning to the first step in which the OOB observations were run through the tree model from which they were excluded, the random forest model records the pairwise frequency for OOB observations that simultaneously end up in the same terminal node [88,95]. These pairwise frequencies are aggregated over all trees and represented within an N × N matrix after dividing by the total number of trees. The elements of this proximity matrix, when subtracted from one correspond to squared Euclidean distances [88]. Using metric multidimensional scaling, it is possible to express visually the feature-based proximity among the regions in our data set (as opposed to proximity across geographical space) in a lower dimensional space. These proximities are visualised in two-dimensional and three-dimensional plots in Figures 9 and 10, respectively. Regions that are plotted closer to each other have a higher degree of similarity with respect to the predictors, and each arm roughly represents a separate class of regions. Darker coloured region names (and circles in the 3-D version) are those with higher per capita patent applications, and larger circle sizes in Figure 10 represent higher levels of work-time flexibility.

Stochastic Gradient Boosting Machine
Arguably, among the models that we apply, the "learning" aspect stands out most prominently in a Gradient Boosting Machine (GBM) developed by Friedman [105]. This is due to the GBM's sequential manner of updating the information learned from each preceding tree. The procedure begins with an initial guess for per capita patents-its sample mean (ȳ)-which is the solution to the minimisation of a squared-error loss function. In the sequential framework, the prediction for the European region i made by the preceding tree is denoted asŷ i,t−1 , where t is the index value for trees (t = 1, . . . , T). The residual r it which corresponds to observation i from the tree t can be expressed as the negative gradient of the loss function, with respect to the prediction evaluated at eachŷ i,t−1 [106]. Fitting a new regression tree to the residuals r it yields terminal regions R mt , each generating an average value γ mt (again, a result obtained through the minimisation of a squared-error loss function [95,105]). Therefore, the tree aims to identify those parts of the data where its preceding counterpart made poor predictions [92]. Each preceding prediction is then shifted towards the true value by a step size. The step size for an observation i belonging to region R mt is equal to its predicted residual γ mt weighted by the shrinkage parameter, also called the learning rate, denoted by ν. The learning rate specifies how much each tree learns from previous errors and 0 < ν < 1. Therefore, if, for instance, a model yields no residuals for a certain group of observations, then there is nothing left for the next model to improve upon the previous predictions. However, if such a claim is false, the subsequent trees can correct this error. Finally, for all i, the gradient boosting algorithm takes the following recursive form:ŷ it =ŷ i,t−1 + νγ mt 1(i ∈ R mt ) (4) and it is iterated until additional trees no longer improve on preceding predictions [95,105]. For any given observation, the serial mending of the predictions improves the model by rectifying any damage caused by possible missteps taken by the previous tree [107]. Furthermore, the slowdown in the incremental approach introduced by the learning rate provides an opportunity for numerous tree models with various structures to successively tackle model errors [92]. Apart from the learning rate, the other parameters that need to be tuned for a GBM are the number of trees and the interaction depth, which is the number of splits for all trees [92]. A smaller learning rate necessitates the usage of a higher number of trees, and an interaction depth of between 4 and 8 is shown to generally perform well [95,105].
A variant of GBM is a Stochastic Gradient Boosting model (S-GBM). In S-GBM, stochasticity is included into the boosting process by randomly subsampling-without replacement-a portion of the training data at each iteration [106]. Further randomisation can be incorporated into the selection of the split variables at each node in each tree, as in a random forest model. Friedman [106] showed that, in small samples, a learning rate of 0.005 and an optimum number of terminal nodes of seven minimise the estimated error, and a sampling fraction of 0.5 alleviates the consequences of over-fitting that may occur due to large tree sizes. Using the parameter values specified above, we apply the GBM and S-GBM methods-with the sampling fraction being one for the former. We determine the sampling fraction for the feature set at each split as 0.8 for the S-GBM procedure and use a minimum number of five observations in a leaf, with the number of trees being equal to 865 and 324 for GBM and S-GBM, respectively (to complement the suggestions of Friedman [106], the remaining tuning parameters have been determined through a grid search using 10-fold cross validation. Taking a grid search approach for all parameters for GBM and S-GBM is computationally and highly challenging. Thorough grid search-based models may have to run for days using standard personal computers). When spatial and country variables are reintroduced, the RMSE of the GBM is 138.4, and the S-GBM RMSE is 132.5. Considering the small number of trees, a manual tweak of the learning rate to 0.1 lowers the latter figure to 125.9.

GBM and S-GBM Results
The variable importances based on the GBM and S-GBM are presented in Figures 11  and 12. Both methods rank the variable Flexible as the predictor with the highest importance, as in the bootstrap aggregation and random forest results. The share of private sector employment, suburban residence, income level, and the job-related variables TimeOff and NeverStress appear in the upper half of the list in both models. The prominence of the effects of job flexibility and private businesses in our results prompts us to take a closer look at these predictors. The centred individual conditional expectation plot (ICE plot; [108]) in Figure 13 plots the predicted per capita patents versus the variable Flexible, holding all other predictors constant. The partial dependence plot (PDP; [105]) is extended to show the effects of both Flexible and Private, as shown in Figure 14. The average effect of job flexibility is represented by the red line in Figure 13, and the figure suggests that some individual observations demonstrate a differing pattern. The deciles of the variable Flexible are shown by the vertical lines just above the x-axis. While the ICE plot provides further understanding of the relationship, it is important to note that assuming away the correlation between Flexible and the remaining predictorson which the partial dependence calculation relies-is quite unrealistic, and the figure should not be over-elaborated.
In Figure 14, the values of the predicted per capita patents are distinguished by different colours, with light yellow representing the highest values. The figure suggests that a higher presence of the private sector in a region, coupled with a higher ratio of individuals enjoying job flexibility, may result in greater regional innovativeness.
In summary, the ML techniques employed in this section have uncovered very useful information in addition to the results from the econometric estimations presented earlier. Particularly, the role of job flexibility, the effect of a suburban environment, freedom to organise one's own work, and the ability to take time off are all identified as important predictors of innovativeness.

Conclusions
Innovation is not a top-down process that can be forced. It needs a facilitating and encouraging environment and an open and unconstrained mindset that has to operate in a comfort zone. Against this background, the present study has made an attempt to identify factors that induce individual and group creativeness, such as a flexible work rhythm, freedom to make choices on the work floor, and the like, by specifically examining modern-age work practices just prior to the introduction of advanced smartphones and tablets, which has confounded the concept of flexibility in the subsequent years.
Armbruster [109] had argued that metacognition -the topmost cognitive process that regulates, oversees, and orchestrates cognition-is highly unconscious and that attempting to control and guide individual creativity in too early stages will have a hampering effect. In the same spirit, our findings clearly suggest that innovativeness requires some individual freedom and a flexible work environment. We have observed in our study that an ideal degree of freedom, complemented by a sense of security and a supporting environment, enhances individual and, therefore regional, innovativeness. As a result, our study has extended and further solidified the theoretical relationship between creativity (in the form of regional innovative output) and personal autonomy by examining different dimensions of individual work-related freedom, i.e., the length of work hours, job flexibility, job autonomy, and ability to take time off. Coupled with other individual level factors arising from one's work such as the levels of stress and the perception of job security, alongside attributes such as the living environment, personal income, and level of education, we have generated new scientific information that can contribute to the extension of the theoretical framework in relation to knowledge production. On the firm level, we have observed the positive role of private businesses. Our results can potentially be added to the pool of evidence together with other research results and influence the labor market side of technology and innovation policies by having an impact on regulations regarding work hours, flexibility, usage of permanent contracts, and support for private establishments.
Undoubtedly, the question of the effect of a relaxed, although responsible, lifestyle at work is too complex to be answered by only one study. Economic, sociological, psychological, and historical viewpoints may need to come together to arrive at a clearer understanding of the individual-level effects of work and place-based geographical factors on innovativeness. There is a clear need for more motivational and cognitive-oriented studies. The current time is characterized by a surge in the availability of Big Data, often in combination with personal perceptions and behavioural stimuli. Against this background, machine learning can serve as a novel and common interdisciplinary methodology for researchers and policy makers in order to understand the individual drivers of creativity and innovation.
Finally, it is remarkable that Elsbach and Hargadon [110] used an analogy from music in their discussion on individual creativity with respect to work design by quoting the jazz musician, Miles Davis, who recognised that the "qualities of musical pieces are not captured in the arrangement of the notes, but also in the arrangement of the silences between notes." An interesting lesson might be that creative minds should not be put in a position in which they deteriorate due to obstructing or frustrating work and discouraging locational influences. Individual innovativeness should be revived and released, enabling firms, regions, and economies to reach higher levels of welfare. Funding: Karima Kourtit and Peter Nijkamp gratefully acknowledge the grant from the Romanian Ministry of Research and Innovation, CNCS-UEFISCDI, project number PN-III-P4-ID-PCCF-2016-0166, within the PNCDI III project ReGrowEU-Advancing ground-breaking research in regional growth and development theories through a resilience approach: towards a convergent, balanced and sustainable European Union (Iasi, Romania).