Next Article in Journal
GPU Accelerated Nonlinear Electronic Circuits Solver for Transient Simulation of Systems with Large Number of Components
Next Article in Special Issue
For Preventative Automated Driving System (PADS): Traffic Accident Context Analysis Based on Deep Neural Networks
Previous Article in Journal
A Path Planning Strategy with Ant Colony Algorithm for Series Connected Batteries
Previous Article in Special Issue
Context-Aware Link Embedding with Reachability and Flow Centrality Analysis for Accurate Speed Prediction for Large-Scale Traffic Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Application of Reinforced Learning-Based Dynamic Pricing for Improvement of Ridesharing Platform Service in Seoul

1
Research Institute of Science and Technology, Hongik University, Seoul 04066, Korea
2
Department of Smart City, Hongik University, Seoul 04066, Korea
3
Department of Urban Planning, Hongik University, Seoul 04066, Korea
*
Author to whom correspondence should be addressed.
Electronics 2020, 9(11), 1818; https://doi.org/10.3390/electronics9111818
Submission received: 29 September 2020 / Revised: 22 October 2020 / Accepted: 30 October 2020 / Published: 2 November 2020
(This article belongs to the Special Issue AI-Based Transportation Planning and Operation)

Abstract

:
As ridesharing services (including taxi) are often run by private companies, profitability is the top priority in operation. This leads to an increase in the driver’s refusal to take passengers to areas with low demand where they will have difficulties finding subsequent passengers, causing problems such as an extended waiting time when hailing a vehicle for passengers bound for these regions. The study used Seoul’s taxi data to find appropriate surge rates of ridesharing services between 10:00 p.m. and 4:00 a.m. by region using a reinforcement learning algorithm to resolve this problem during the worst time period. In reinforcement learning, the outcome of centrality analysis was applied as a weight affecting drivers’ destination choice probability. Furthermore, the reward function used in the learning was adjusted according to whether the passenger waiting time value was applied or not. The profit was used for reward value. By using a negative reward for the passenger waiting time, the study was able to identify a more appropriate surge level. Across the region, the surge averaged a value of 1.6. To be more specific, those located on the outskirts of the city and in residential areas showed a higher surge, while central areas had a lower surge. Due to this different surge, a driver’s refusal to take passengers can be lessened and the passenger waiting time can be shortened. The supply of ridesharing services in low-demand regions can be increased by as much as 7.5%, allowing regional equity problems related to ridesharing services in Seoul to be reduced to a greater extent.

1. Introduction

Since the 1960s, with the beginning of industrialization, Korea experienced a rapid urbanization [1,2,3]. As such, the transportation infrastructure has also developed steadily, greatly improving the mobility of citizens. However, in metropolitan areas, such as the capital city and its surrounding regions, the expansion of urban areas and overpopulation caused increased traffic and consequently other social problems, including traffic congestion and air pollution. To minimize such social costs, various policies to facilitate public transportation and taxis have been actively promoted [4]. Despite the efforts, the current services are short in providing an equal level of mobility to all residents, requiring enhancement. In particular, in Seoul Metropolitan City, despite its excellent public transportation infrastructure, there is a constant inconvenience of users with a need for enhancing problems caused by congestion during rush hours and a shortage of ridesharing (including taxi) supply late at night on the outskirts of the city and in hilly areas [5].
Against this backdrop and with the recent introduction of the mobility as a service(MaaS) concept, efforts have been made to improve individuals’ mobility and accessibility through the integrated use of shared and cutting-edge modes of transportation [6,7,8]. MaaS consists of various means of transportation, including public transportation and a mobile platform-based ridesharing service. However, the ridesharing service is usually owned by the private sector, oriented toward profitability. Therefore, the provision of such a service tends to be concentrated in areas with high demand. Even if the ridesharing service were to expand, those users living in areas with low demand, especially suburban areas, are expected to face inconvenient services, such as longer waiting times due to the lower availability of vehicles. Especially during late-night hours when public transportation no longer operates, there are only limited options available, thus leading to further exacerbation of the imbalance between supply and demand. Therefore, a solution for this problem is needed. In a previous study [9] on experiences with taxi service refusal, respondents answered that they had difficulties getting taxi services or were refused during late-night hours (48.7%) when their destinations were either remote areas or outside of the city boundary (32.8%). From the perspective of users, this is recognized as a serious problem. The purpose of this study was to develop a dynamic pricing scheme to attract ridesharing (including taxi) drivers to mobility-disadvantaged regions during late-night hours when public transportation services are either reduced or come to a halt.
In the meantime, with the recent Fourth Industrial Revolution, the overall social paradigm has changed. A hyperintelligent society based on cutting-edge technologies, such as artificial intelligence, is being implemented. In the transportation sector, studies related to the introduction of big data- and artificial intelligence-based technology are growing [10]. Indeed, overseas ridesharing platforms, namely, Uber and Lyft, optimize supply and demand, as well as profitability, through AI-based dynamic pricing to retain the number of vehicles needed to respond to high demand in certain regions. Among the various subcategories of artificial intelligence, reinforcement learning has advantages in exploring unknown areas and identifying optimal outcomes through repeated exploration, unlike supervised learning, which is based on already available data. In contrast to other countries, there are not enough data related to ridesharing services in Korea and, accordingly, reinforcement learning would be more suitable. Consequently, this study conducted a reinforcement learning method using Seoul’s taxi data to determine regionally appropriate levels of dynamic ridesharing fare rates with the purpose of improving the supply of ridesharing services in the mobility-disadvantaged regions in Seoul.
The paper is organized as follows: Section 2 reviews studies on the spatial fairness issues of transportation services and dynamic pricing applications for ridesharing services. Section 3 discusses the analysis framework and reinforcement learning method applied in this study. Subsequently, the results of the analysis are presented and discussed in Section 4, while the conclusions are drawn in Section 5.

2. Literature Review

2.1. Studies Analyzing Spatially Marginalized Areas in Terms of Transportation Services

In their study, Lee et al. [11] identified the mobility of different marginalized groups using smart card data and evaluated the mobility of groups highly reliant on public transportation. Building upon this, the study also categorized regions into different types according to their need for mobility improvement and concluded that mainly the outskirt areas of a city require improvements. Ha et al. [12] located areas marginalized from public transportation services in Seoul, using real travel characteristics gathered using Google and T map navigation application programming interfaces (APIs), and they analyzed areas with high priority in service enhancement. Data, such as travel characteristics and socioeconomic indices, were analyzed, identifying Gangbuk-gu, Seongbuk-gu, Seodaemun-gu, Jungnang-gu, and the southeast zone as areas requiring improvements in public transportation. Han [13] deduced spatially marginalized areas by assessing user mobility and the service level of providers. Moreover, the study reviewed the potential equity issue caused among different groups and suggested ways to improve it. According to the study, overall public transportation showed a satisfactory level from the users’ perspective; however, there was a large deviation in terms of the supply level by region. Notably, on the outskirts of the city, including Nowon-gu and Gwanak-gu, the gap between supply and demand appeared to be wide and, thus, most urgently requiring an improvement in supply.
From an accessibility perspective, Lee et al. [14] developed various indices for the connectivity, directness, and diversity of public transportation using transportation card data and assessed each transportation zone. This study confirmed that, with a higher connectivity of public transportation, it had more routes and better directness. On the contrary, in zones located on the outskirts of the city, marginalization from public transportation was vivid. Kim et al. [15] and Yoon et al. [16] studied regional marginalization and inequity by considering not just spatial accessibility, but also social classes. Kim et al. overlaid and compared socioeconomic characteristics and city zones using location data of public transportation in Daegu. According to their analysis, the low accessibility and environmental inequity of socially disadvantaged people (the aged, recipients of national basic livelihood benefits, etc.) in suburban areas were confirmed. Yoon et al. calculated the inequity index of socially marginalized people on the basis of a Gini-style index and the methods of accessibility measure developed by Curie. As for public transportation accessibility, the regional gap was bigger for the subway than the bus. When this was overlaid on top of the data for the socially disadvantaged group, inequity was confirmed to be greater for the subway than the bus.
There were studies assessing equity depending on regional differences in infrastructure. Kim et al. [17] analyzed disadvantaged regions by overlaying service areas of public transportation and confirmed that suburb areas were mainly in a disadvantageous position. Furthermore, when considering socioeconomic characteristics, the study concluded that there was a gap in public transportation infrastructure among regions. Lee et al. [18] used Seoul Metropolitan Household Travel Survey data to measure regional equity among different income brackets depending on the levels of transportation infrastructure. The study showed that lower spatial equity led to longer total traveling time. Bin et al. [19] carried out a spatial cluster analysis at the administrative unit (Eup, Myeon, and Dong) level with transportation infrastructure indicators and travel behavior to assess equity in Gyeonggi-do province (excluding Seoul Metropolitan City and Incheon City). The results clearly showed gaps between areas closer to Seoul and those on the outskirts of the city. In particular, equity at the infrastructure level in northern Gyeonggi-do province was low.

2.2. Dynamic Pricing Studies on Ridesharing Service

Before reviewing previous literature, dynamic pricing can be defined as a strategy in which prices change flexibly for the same product or service depending on the market situation [20,21,22]. This strategy is mainly employed with respect to e-commerce, flight tickets, and hotel booking and demand management. This results in the optimization of selling products and services in an environment where the price can be easily adjusted. With regard to dynamic pricing for ridesharing services, there were studies on profitability improvement and the determination of an appropriate price through a pricing strategy [23,24,25,26,27,28,29,30], studies analyzing the elements of a pricing strategy affecting customers or drivers [31,32,33], and a study based on reinforcement learning [34].
First, Banerjee et al. [23] validated the performance of dynamic pricing by suggesting a ridesharing model considering two aspects: the stochastic dynamics of the market and the strategic decisions of the drivers, passengers, and platform. According to the analysis, depending on supply and demand conditions, flexible pricing resulted in increased total utility. Zeng et al. [24] researched the dynamic pricing strategy in accordance with potential users considering the destination of taxies. Markov Decision Process (MDP) was established by considering the cost of pick-ups at the destination. The total utility was enhanced compared to the fixed cost case. Hall et al. [25] studied the economic utility of surge pricing by analyzing Uber data.
When prices rise in line with the surge price algorithm, the delta affected the improving profitability of drivers, supply, and efficiency. Moreover, if surge pricing was not employed during peak hours, passenger waiting time was extended as drivers did not pick up passengers and passenger utility dropped.
In an analysis of the effects of dynamic pricing on drivers through Uber cases, Chen et al. identified that the surge price has a negative impact on passengers and a positive impact on drivers. Furthermore, the study found unfairness along the regional border of surge pricing. Chen et al. [32] studied changes in the number of Uber drivers depending on changes in surge price. The analysis confirmed that there was a higher rate of vehicle operation, as well as changes in operating hours, during the time period in which higher profit was expected from the drivers’ perspective. Kooti et al. [33] analyzed the impact of dynamic pricing and income on the behavior of drivers, and they found that drivers operating vehicles during peak hours earned a higher income compared to nonpeak-hour drivers.
Wu et al. [34] simulated the application of dynamic pricing to ridesharing services. They compared four pricing methodologies, namely, (1) statistic pricing, (2) proportional pricing, (3) batch updates, and (4) reinforcement learning with the goal of profit maximization in a single Origin-Destination(OD) scenario. The simulation found that pricing based on reinforcement learning increased the total profit and individual profit of drivers the most.

2.3. Summary of the Review

Previous studies on mobility-disadvantaged regions mainly focused on evaluating areas marginalized from public transportation from a user’s perspective utilizing big data, such as smart card data, household travel surveys, and Geographic Information System Data Base (GIS DB: location data for bus and subway, thematic transportation map, etc.). These studies identified that public transportation services in these regions were not adequately provided. On the other hand, they paid little attention to the late-night mobility inconvenience in regions after public transportation services had ended.
As for dynamic pricing in ridesharing services, studies were carried out mainly using Uber cases. As Uber discloses operation data, several studies related to dynamic pricing were conducted. In particular, the surge-based pricing analysis of Uber showed that it helped increase driver supply and operational profit. On the contrary, if a single-fare system was applied without flexibility, overall utility decreased due to a lower rate of matching and a longer waiting time when suppliers decided not to take passengers. However, there were no studies measuring regional differences in dynamic pricing; thus, it is necessary to determine ways of addressing drivers’ refusal to take passengers when taking regional differences into account.
Accordingly, a study evaluating the dynamic pricing solution is required to improve the quality of mobility services in disadvantaged regions in Seoul.

3. Analysis Methodology

To simulate dynamic pricing using reinforcement learning, this study first conducted a centrality (in-bound and out-bound) analysis with a district(gu)-level OD matrix to develop a regional indicator of degree centrality to be used in reinforcement learning analysis. Then, the indicators were applied to a reinforcement learning simulation.

3.1. Analysis Methodology

3.1.1. Degree Centrality Analysis Methodology

Degree centrality is an indicator showing how many nodes are connected to a certain node. In social network theory, degree centrality is defined as the number of nodes linked directly to any given node. This study measured the out-degree and in-degree of each region (node) from a centrality perspective depending on the level of node connectivity, using them as indicators to be applied in reinforcement learning. Centrality can be calculated as follows:
C D ( N i )   =   j   =   1 g x i j g 1 ,   i j ,
where C D is the standardized degree centrality of node i , j   =   1 g x i j is the degree centrality of node j, and g is the number of nodes. If there is no direction in the network, the above equation simply shows the degree of nodes. On the other hand, if directions are present, the equation distinguishes out-degree centrality ( C o u t D ) and in-degree centrality ( C i n D ). Here, out-degree centrality is defined as the level of connections going out from a certain node to other nodes, and in-degree centrality is defined as the level of connections coming in from other nodes to a certain node. In this study, in-degree centrality referred to the number of vehicles coming into a certain zone (Gu), thus indicating a region as a frequently selected destination by passengers (usually residential areas during late-night periods). On the contrary, out-degree centrality referred to the number of vehicles traveling out of a certain zone, thus indicating a region as a preferred destination by drivers (central and subcentral regions). Therefore, this study identified indicators by region considering both out-degree and in-degree centrality.

3.1.2. Reinforcement Learning Methodology

Reinforcement learning is a learning method through which an agent chooses an action to take in an environment to maximize reward. This affects not only the immediate reward due to the action of the agent, but also the long-term reward. The main characteristics of reinforcement learning are a trial-and-error search and delayed reward (Figure 1). This study used the Q-learning algorithm for analysis.
In the Q-learning algorithm [35,36], Q initially has an arbitrary fixed value. When the agent selects an action ( a ) in accordance with a learning step ( t ), an immediate reward ( r ) is observed before entering a new state ( s ), updating the Q value. The key characteristic of this algorithm is the value iteration update, using the weighted average of the old value and new information. The equation of Q-learning can be expressed as follows:
Q ( s t ,   a t ) ( 1 α ) Q ( s t ,   a t ) + α ( r t + γ max α Q ( s t + 1 ,   a ) ,
where α is the learning rate which updates the current Q value with immediate reward and/or future expected value, usually between 0 and 1. When the value is 0, the current Q value is continuously used without any update according to learning. On the contrary, if the value is 1, the previous Q value is ignored and updated automatically. γ is a discount factor which is a variable explaining the difference between immediate and future rewards, also between 0 and 1. When the value is 0, the agent takes a myopic action as by making an update with immediate reward. When the value is 1, the agent leans toward a future reward, underestimating the current reward. In this study, on the basis of existing studies, we experimentally set the learning rate to 0.1 and discount factor to 0.9.

3.2. Environment Setting for Reinforcement Learning Simulation

3.2.1. Q-Table Configuration

To carry out the reinforcement learning simulation using the Q-learning algorithm, a Q-table needed to be defined through state and action. The Q-table is a matrix that compiles rewards for all states and actions, and it is updated after each learning step. Here, 600 OD cells were used in an OD matrix of 25 zones in Seoul excluding intrazonal travel. The state refers to the price that can potentially be generated in each OD cell depending on time slots, which can be expressed as 13 (price range) × 6 h (time round, Tr) for each OD cell. In total, there were 13 actions, including a standard price and surge prices from 0.6-fold to 3.0-fold the standard price with an interval of 0.2. The composition of the district OD matrix is shown in Figure 2, with a separate layer constructed for each cell’s travel time and passage cost ( P b a s e ) to be used for analysis.

3.2.2. Action Model

The simulation included passengers and drivers for each cell in the OD matrix and determined whether or not a match was feasible on the basis of the choice probability of passengers and drivers. Applying the method used by Wu et al. [34], passengers selected a price according to the surge according to a normal distribution. On the other hand, drivers applied choice probability according to a normal distribution with the weight of the centrality analysis outcome in accordance with their destination. Centrality analysis was applied to taxi operation data on the basis of 150 M links provided for Seoul Metropolitan City. The analysis was conducted using data from April to September of 2017 which were processed to generate OD data for each district (gu).
For the choice probability model of drivers and passengers, a surge of 1 (base fare of taxi) was applied by consulting the result of National Bureau of Economic Research(NBER, 2016) study [37]. In addition, the probability followed a cumulative normal distribution using the following equation:
  Φ ( x )   =     x N ( 0 , 1 ) ( x ) d x .
The choice probability models of passengers (Equation (4)) and drivers (Equation (5)) according to the surge are expressed below.
P ( a c c e p t | P o f f e r )   =     Φ ( P o f f e r P b a s e σ ) × W 1 ,
P ( a c c e p t | P o f f e r )   =     Φ ( P b a s e P o f f e r σ ) × W 2 +   ( T r C o u t D ( N j ) T r C i n D ( N j ) ) × W 3 ,
where W 1 is the weight modifying the choice probability of passengers when surge = 1, W 2 is the weight modifying the choice probability of drivers when surge = 1, T r C o u t D ( N j ) is the out-degree centrality value of a specific time round ( T r ), T r C i n D ( N j ) is the in-degree centrality value of a specific time round ( T r ), and W 3 is the weight of the equity index allowing differentiation by region.

3.2.3. Reward Function

The criteria for reward were different depending on matching. Once a driver was matched with a passenger, the reward was the base fare ( P f a r e ) for the travel between origin and destination multiplied by the surcharge coefficient ( S a ). If a driver was not matched with a passenger, a negative reward was applied in line with the waiting time value ( W t ). The waiting time was set to 8 min considering the fact that the average waiting time during late-night periods when a surcharge is applied is 8.1 min according to a study in Seoul Metropolitan City [38]. The time value ( V t ) was calculated on the basis of the taxi fare for 1 min in the OD matrix. While it increased profitability through a surcharge, it was also designed to provide a negative reward when unmatched by multiplying the time value for 8 min by the surcharge. The equation for the reward can be expressed as shown in Equation (6). Learning was conducted separately for when a negative value was applied to waiting time and for when it was not. Through a comparison, the study identified the appropriate reward function.
r =   { m a t c h e d P f a r e ×   S a                                                     u n m a t c h e d { W t × V t × S a   a l t 1 )     0                                             a l t 2 ) .

3.2.4. Q-Learning Algorithm

A total of 20,000 Q-learning reiterations were conducted by taking the index value for each zone into account in the late-night period (Figure 3). In one episode, the goal was to identify the optimal surge for each time slot during 6 h of operation for each OD matrix. The Q-table included states (current state and price) for each time slot and action value by applying the concept of time. The travel time and price required to calculate the reward value (operating profit) for each state referred to a separate OD table. The weight values applied to the preference of drivers were the centrality indices for each time slot and zone. The reward function was created separately depending on the application of the time value.

4. Analysis Results

4.1. The Analysis Result of Areas Expected to Be Marginalized in Ridesharing Services

In order to identify the indices to be applied to reinforcement learning, the taxi operation data in Seoul were categorized into different zones according to 25 districts (gu), and the OD matrix was formed on the basis of the volume of pick-ups and drop-offs using district codes. Consequently, the study analyzed centrality and tried to identify a standardized index for each zone to be applied to reinforcement learning. The origin (the number of pick-ups) was compiled by matching administrative districts through the GIS spatial join function, while the destination (the number of drop-offs) was compiled by using the district(gu) code information for each vehicle departure. Any data including areas surrounding Seoul were excluded.
Using the OD matrix, degree centrality analysis for each zone was conducted, and the average in-degree and out-degree centrality was calculated for weekdays and weekends from 10:00 p.m. to 4:00 a.m. (Figure 4 and Figure 5). According to the analysis, in-degree centrality displayed similar patterns for both weekends and weekdays, whereas the regional deviation was smaller than seen for out-degree centrality. Out-degree centrality was high in the Central Business Districts(CBDs) on weekdays and similarly high on weekends.
Areas located on the outskirts of the city had a lower degree centrality. In particular, outskirt areas showed lower out-degree centrality on weekdays. Residential areas on the outskirts of Seoul showed higher in-degree centrality. On the other hand, central and subcentral regions of the city showed higher out-degree centrality (Figure 6). The patterns were similar to those found in previous studies showing the concentration of taxi pick-ups during late-night hours near CBDs [39]. There were large differences in out-degree centrality by zone before and after midnight when most public transportation services were suspended, with the difference growing as the night went on.

4.2. Results of Reinforcement Learning Simulation

The algorithm presented in Figure 3 was used in the learning of 600 OD cells. When the reward function was defined by a simple matching rate, there was a limitation in comparison among regions as the fare and travel time for each OD were different. Therefore, the simulation implemented a reward value considering travel time and cost. Moreover, analysis was conducted depending on the application of a negative reward value in the unmatched condition. The matching rate for each learning step when a waiting time with a negative reward was applied (alt1) or not (alt2) is depicted in Figure 7.
When the negative reward was considered, the matching rate increased before converging as learning was reiterated. When it was not considered, the matching rate decreased before converging. According to this result, it can be concluded that the application of a negative reward was necessary when the reward function used the travel cost but not the matching rate. However, the current negative reward for the waiting time used a fixed value from a previous study; thus, for real-world applications, the expected waiting time according to a driver’s choice probability should be calculated for each OD.
Using the deduced average surge, when a fare was calculated for each OD, the distribution could be expressed as shown in Figure 8, where the x-axis is the travel distance between origin and destination, and the y-axis is the price. Then, price levels were compared in terms of the base fare, base fare with late-night surcharge (additional 20% to base fare), late-night surcharge + ride-hail fee (KRW 3000, the highest fee charged by existing platform companies), and surge price (the outcome of learning for each OD). For travel distances shorter than 10 km, the differences among fares were not severe. However, the gap grew wider when the distance increased (see Figure 8a). This is because the late-night surcharge + ride-hail fee (yellow dotted line) was a flat fare regardless of traveling distance, while the surge fare (blue dotted line) was determined using a certain ratio, increasing the absolute price with traveling distance. Table 1 displays the three average prices for different distances, showing the price gap widening as the distance increased.

4.3. Changes in Centrality after Surge-Driven Supply Increase

Centrality analysis was carried out after recalculating the traffic volume in accordance with the surge previously identified for each OD matrix and time slot. The traffic volume was recalculated as follows:
N v o l i , j   =   ( O v o l i , j ×   S i , j ) × i , j O v o l i , j i , j O v o l i , j × S i , j ,
where i , j are the origin and destination, N v o l i , j is the recalculated traffic volume, O v o l i , j is the previous traffic volume, S i , j is the optimal surge for travel (optimal surge for different OD as identified from reinforcement learning), and i , j O v o l i , j is the sum of traffic volume. In centrality analysis, the indicator evaluating the volume of vehicles traveling to a certain zone was the in-degree centrality. When in-degree centrality increases, it can be said that supply is increasing toward that region.
According to the level of improved spatial equity by region, in-degree centrality decreased in the central and subcentral regions of the city, such as Gangnam-gu and Jongno-gu, while the in-degree centrality increased in residential areas on the outskirts of the city, such as Songpa-gu, Gangseo-gu, Dongjak-gu, and Nowon-gu (Figure 9). When this was overlaid on top of the previous hotspot analysis, it was found that the in-degree centrality was drastically improved in districts (gu) located on the outskirts of Seoul Metropolitan City where the number of drop-offs or vacant vehicles was greater compared to the number of pick-ups (Figure 10b). As a result, it can be expected that vehicle supply would be enhanced by as much as 7.5% when applying a higher surge to the region predicted to have a lower choice probability from the perspective of drivers. At the same time, equity in service supply would be improved by reducing the waiting time of passengers in marginalized regions with low taxi demand.

5. Conclusions

As ridesharing (including taxi) services are often run by private companies, profitability is the top priority in operation. This leads to an increase in drivers’ refusal to take passengers to areas with low demand where they have difficulties finding subsequent passengers, causing problems such as extended waiting time when hailing a vehicle for passengers bound for these regions. This problem differs depending on time and region. In late-night hours and in suburban regions, the imbalance between supply and demand is especially widened, worsening the problem. In order to address this problem, solutions were proposed in this study through a dynamic pricing strategy using reinforcement learning algorithms.
This study used Seoul city’s taxi data to find appropriate fare surge rates for ridesharing services between 10:00 p.m. and 4:00 a.m. In reinforcement learning, the outcome of centrality analysis was applied as the weight affecting drivers’ destination choice probability. Moreover, the reward function used during learning was adjusted according to whether or not a passenger waiting time value was applied. Profit was used as the reward value. By applying a negative reward for the passenger’s waiting time, a more appropriate surge fare level could be identified. Across the region, the average surge level amounted to 1.6. Regions located on the outskirts of the city in predominantly residential regions such as Gangdong-gu, Dongjak-gu, Eunpyeong-gu, and Gangseo-gu showed a higher surge. On the contrary, central areas, such as Gangnam-gu, Jongno-gu, and Jung-gu, had a lower surge. The findings showed that the supply of ridesharing services in low-demand regions could be increased by as much as 7.5% using surge fares, thereby reducing regional equity problems related to ridesharing services in Seoul to a great extent.
This study conducted a reinforcement learning-based dynamic pricing simulation to respond to the regional equity problem of ridesharing (including taxi) services in Seoul. A novel approach was presented using dynamic pricing as a way to mitigate the spatial equity problem by affecting ridesharing supply, unlike most previous dynamic pricing studies which simply targeted higher profitability. Notably, it was shown that a surge rate change in fares could reduce the indirect refusal of drivers to take passengers to unpreferred areas. With additional real-time ridesharing user data, the Deep Q-Network(DQN) technique can be adopted to conduct a smaller-scale spatial analysis of ridesharing services. Furthermore, with more knowledge on fare sensitivity by user group, the dynamic pricing approach proposed in this study can significantly contribute to resolving the spatial equity problem in mobility services in the future.

Author Contributions

Conceptualization, K.Y.H.; data curation, J.S.; investigation, Y.J.C. and M.H.K.; methodology, J.S.; project administration, M.H.K.; software, J.S.; writing—original draft, J.S.; writing—review and editing, K.Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a grant (20TLRP-B148970-03) from the Transportation and Logistics R&D Program funded by the Ministry of Land, Infrastructure, and Transportation of the Korean government.

Acknowledgments

This paper is a modification and expansion of Jae In Song’s doctoral dissertation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Park, J.; Kim, D.; Ko, Y.; Kim, Y.; Park, K. Urbanization and urban policies in Korea. Korea Res. Inst. Hum. Settl. 2010, 2010, 56. [Google Scholar]
  2. CHO, M. Trend and prospect of urbanization in Korea: Reflections on Korean cities. Econ. Soc. 2003, 60, 10–39. [Google Scholar]
  3. Lim, K. Characteristics of Urbanization and its Policies in Korea. J. Korean Reg. Dev. Assoc. 2004, 16, 139–156. [Google Scholar]
  4. Hwang, S. Revitalizing the Transportation Market with the Public Transportation Promotion Act. Korean Soc. Transp. 2005, 2, 7–13. [Google Scholar]
  5. Hong, S.; Yoon, H.; Yu, K.; Yun, J. A Study on Deriving of Vulnerable Sections in Public Transportation Services. Seoul Inst. 2019, 2019, BR-02. [Google Scholar]
  6. Heikkilä, S. Mobility as a Service—A proposal for Action for the Public Administration, Case Helsinki. Master’s Thesis, Aalto University, Aalto, Finland, 2014. [Google Scholar]
  7. Cox, N.C.J. Estimating Demand for New Modes of Transportation Using a Context-Aware Stated Preference Survey. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 2015. [Google Scholar]
  8. Finger, M.; Bert, N.; Kupfer, D. Mobility-as-a-SERVICES: From the Helsinki Experience to a European Model? European Transport; Regulation Observer: San Domenico di Fiesole, Italy, 2015. [Google Scholar]
  9. Lee, J.; Kim, D. The role of smart mobility in resolving the imbalance in transportation supply and demand. In Proceedings of the 2018 Spring Academic Conference of the Korea Society of Management Information System, Seoul, Korea, 10 November 2018; pp. 139–158. [Google Scholar]
  10. Kim, K.; Kim, T.; Cho, S.; Jang, H. Deregulation Strategies for Promoting Innovation of Technology and Services in Transport Sector in the Era of 4th Industrial Revolution; The Korea Transport Institute: Sejong City, Korea, 2019. [Google Scholar]
  11. Lee, H.; Ha, J.; Lee, S. An Analysis on the Equity of Public Transit Service using Smart Card Data in Seoul, Korea—Focused on the Mobility of the Disadvantaged Population Groups. J. Korean Reg. Sci. Assoc. 2017, 33, 101–113. [Google Scholar]
  12. Ha, J.; Lee, S. An Analysis of Vulnerable Areas for Public Transit Services using API Route Guide Information—Focused on the Mobility to Major Employment Centers in Seoul, Korea. J. Korea Plan. Assoc. 2016, 51, 163–181. [Google Scholar] [CrossRef]
  13. Han, D. The Analysis of Public Transport Accessibility and Equity in Seoul. Ph.D. Thesis, Graduate School of Konkuk University, Seoul, Korea, 2016. [Google Scholar]
  14. Lee, H.; Ko, S.; Kim, D. Development of Equity Indicators for Public Transportation Using Transportation Card Data. In Proceedings of the 2018 Spring Academic Conference of The Korean Institute of Intelligent Transport. Systems, Jeju-do, Korea, 19–21 April 2018; pp. 471–476. [Google Scholar]
  15. Jun, B.; Kim, A. Environmental Equity Analysis of the Accessibility to Public Transportation Services in Daegu City. J. Korean Assoc. Geogr. Inf. Stud. 2012, 15, 76–86. [Google Scholar]
  16. Yun, J.; Woo, M. Empirical Study on Spatial Justice through the Analysis of Transportation Accessibility of Seoul. J. Korea Plan. Assoc. 2015, 50, 69–85. [Google Scholar] [CrossRef]
  17. Kim, J.; Kang, S.; Kwon, J. The Spatial Characteristics of Transit-Poors in Urban Areas. J. Korean Assoc. Geogr. Inf. Stud. 2008, 11, 1–12. [Google Scholar]
  18. Lee, W.; Bin, M.; Moon, J.; Cho, C. Analysis and Policy Proposal on the Vulnerable Areas of Public Transportation Accessibility in Gyeonggi-do. 2012 Acad. Conf. J. Korean Soc. Transp. 2012, 67, 47–50. [Google Scholar]
  19. Bin, M.; Lee, W.; Moon, J.; Joh, C. Integrated Equity Analysis Based on Travel Behavior and Transportation Infrastructure: In Gyeonggi-Do Case. J. Korean Soc. Transp. 2013, 31, 47–57. [Google Scholar] [CrossRef] [Green Version]
  20. Elmaghraby, W.; Keskinocak, P. Dynamic pricing in the presence of inventory considerations: Research overview, current practices, and future directions. Manag. Sci. 2003, 49, 1287–1309. [Google Scholar] [CrossRef]
  21. Rouse, M. What is Dynamic Pricing?—Definition from WhatIs. com. WhatIs. com. 2015. Available online: https://whatis.techtarget.com/definition/dynamic-pricing (accessed on 1 November 2020).
  22. den Boer, A.V. Dynamic pricing and learning: Historical origins, current research, and new directions. Surv. Oper. Res. Manag. Sci. 2015, 20, 1–18. [Google Scholar] [CrossRef] [Green Version]
  23. Banerjee, S.; Johari, R.; Riquelme, C. Dynamic pricing in ridesharing platforms. ACM SIGecom Exch. 2016, 15, 65–70. [Google Scholar] [CrossRef]
  24. Zeng, C.; Oren, N. Dynamic Taxi Pricing; IOS Press: Amsterdam, The Netherlands, 2014. [Google Scholar]
  25. Hall, J.; Kendrick, C.; Nosko, C. The Effects of Uber’s Surge Pricing: A Case Study; The University of Chicago Booth School of Business: Chicago, IL, USA, 2015. [Google Scholar]
  26. Li, S.; Fei, F.; Ruihan, D.; Yu, S.; Dou, W. A dynamic pricing method for carpooling service based on coalitional game analysis. In 2016 IEEE 18th International Conference on High Performance Computing and Communications; IEEE 14th International Conference on Smart City; IEEE 2nd International Conference on Data Science and Systems (HPCC/SmartCity/DSS); IEEE: Piscataway Township, NJ, USA, 2016; pp. 78–85. [Google Scholar]
  27. Waserhole, A.; Jost, V. Pricing in vehicle sharing systems: Optimization in queuing networks with product forms. EURO J. Transp. Logist. 2016, 5, 293–320. [Google Scholar] [CrossRef] [Green Version]
  28. Qiu, H.; Li, R.; Zhao, J. Dynamic pricing in shared mobility on demand service. arXiv 2018, arXiv:1802.03559. Available online: https://arxiv.org/pdf/1802.03559.pdf (accessed on 1 November 2020).
  29. Korolko, N.; Woodard, D.; Yan, C.; Zhu, H. Dynamic Pricing and Matching in Ride-Hailing Platforms. 2018. Available online: https://api.semanticscholar.org/CorpusID:53059883 (accessed on 1 November 2020).
  30. Avdeenko, T.; Khateev, O. Taxi Service Pricing Based on Online Machine Learning. In International Conference on Data Mining and Big Data; Springer: Singapore, 2019; pp. 289–299. [Google Scholar]
  31. Chen, L.; Mislove, A.; Wilson, C. Peeking beneath the hood of uber. In Proceedings of the 2015 Internet Measurement Conference, Tokyo, Japan, 28–30 October 2015; pp. 495–508. [Google Scholar]
  32. Chen, M.K.; Sheldon, M. Dynamic Pricing in a Labor Market: Surge Pricing and Flexible Work on the Uber Platform. UCLA Anderson. 2015. Available online: https://www.anderson.ucla.edu (accessed on 1 November 2020).
  33. Kooti, F.; Grbovic, M.; Aiello, L.M.; Djuric, N.; Radosavljevic, V.; Lerman, K. Analyzing Uber’s ride-sharing economy. In Proceedings of the 26th International Conference on World Wide Web Companion, Perth, Australia, 3–7 April 2017; pp. 574–582. [Google Scholar]
  34. Wu, T.; Joseph, A.D.; Russell, S.J. Automated Pricing Agents in the on-Demand Economy; University of California at Berkeley: Berkeley, CA, USA, 2016; UCB/EECS-2016-57. [Google Scholar]
  35. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  36. Van Otterlo, M.; Wiering, M. Reinforcement learning and markov decision processes. In Reinforcement Learning; Springer: Berlin/Heidelberg, Germany, 2012; pp. 3–42. [Google Scholar]
  37. Cohen, P.; Hahn, R.; Hall, J.; Levitt, S.; Metcalfe, R. Using Big Data to Estimate Consumer Surplus: The Case of Uber. Available online: https://www.nber.org/system/files/working_papers/w22627/w22627.pdf (accessed on 1 November 2020).
  38. An, G.; Kang, T.; Kim, B. Taxi Service Evaluation Service Report 2014; Policy Report 186; Seoul Institute: Seocho-gu, Seoul, Korea, 2015. [Google Scholar]
  39. An, G. Seoul’s Taxi Usage and Operation Status and Improvement Plan; Policy Report 186; Seoul Institute: Seocho-gu, Seoul, Korea, 2015. [Google Scholar]
Figure 1. Conceptual diagram of reinforcement learning.
Figure 1. Conceptual diagram of reinforcement learning.
Electronics 09 01818 g001
Figure 2. OD matrix, where each OD cell is the location data for each OD movement in the matrix.
Figure 2. OD matrix, where each OD cell is the location data for each OD movement in the matrix.
Electronics 09 01818 g002
Figure 3. Q-learning algorithm pseudo code.
Figure 3. Q-learning algorithm pseudo code.
Electronics 09 01818 g003
Figure 4. Results of centrality analysis between 10:00 p.m. and 4:00 a.m. on weekdays: (a) average in-degree centrality; (b) out-degree centrality.
Figure 4. Results of centrality analysis between 10:00 p.m. and 4:00 a.m. on weekdays: (a) average in-degree centrality; (b) out-degree centrality.
Electronics 09 01818 g004
Figure 5. Results of centrality analysis between 10:00 p.m. and 4:00 a.m. on weekends: (a) average in-degree centrality; (b) out-degree centrality.
Figure 5. Results of centrality analysis between 10:00 p.m. and 4:00 a.m. on weekends: (a) average in-degree centrality; (b) out-degree centrality.
Electronics 09 01818 g005
Figure 6. Results of out-degree centrality versus in-degree centrality analysis between 10:00 p.m. and 4:00 p.m.: (a) weekday; (b) weekend.
Figure 6. Results of out-degree centrality versus in-degree centrality analysis between 10:00 p.m. and 4:00 p.m.: (a) weekday; (b) weekend.
Electronics 09 01818 g006
Figure 7. Matching rate comparison depending on time value.
Figure 7. Matching rate comparison depending on time value.
Electronics 09 01818 g007
Figure 8. Price distribution for different OD when a surge was applied (using the average surge for different OD): (a) total range; (b) short distance.
Figure 8. Price distribution for different OD when a surge was applied (using the average surge for different OD): (a) total range; (b) short distance.
Electronics 09 01818 g008
Figure 9. Change in in-degree centrality: (a) before; (b) after.
Figure 9. Change in in-degree centrality: (a) before; (b) after.
Electronics 09 01818 g009
Figure 10. (a) Rate of change in in-degree centrality; (b) comparison between in-degree centrality and hotspot analysis.
Figure 10. (a) Rate of change in in-degree centrality; (b) comparison between in-degree centrality and hotspot analysis.
Electronics 09 01818 g010
Table 1. Traffic volume and average price (KRW) for difference distances.
Table 1. Traffic volume and average price (KRW) for difference distances.
DistanceTraffic VolumeNumber of OD ZonesSurge PriceLate-Night Surcharge 1Late-Night Surcharge + Ride-Hail Fee 1
Shorter than 5 km
5–10 km
10–15 km
15–20 km
20–25 km
Longer than 25 km
28.5%
38.5%
18.4%
8.5%
4.2%
1.8%
46
116
146
108
86
99
9494.6
14,845.0
22,264.0
29,625.4
36,244.8
45,761.9
7017.3
10,450.8
15,846.3
20,863.1
25,203.1
31,731.4
10,017.3
13,450.8
18,846.3
23,863.1
28,203.1
34,731.4
Total100.0%60027,078.119,006.222,006.2
1 The late-night surcharge is surge 1.2 of the base amount and an additional Hail Fee of 3000 KRW (based on the existing platform price).
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Song, J.; Cho, Y.J.; Kang, M.H.; Hwang, K.Y. An Application of Reinforced Learning-Based Dynamic Pricing for Improvement of Ridesharing Platform Service in Seoul. Electronics 2020, 9, 1818. https://doi.org/10.3390/electronics9111818

AMA Style

Song J, Cho YJ, Kang MH, Hwang KY. An Application of Reinforced Learning-Based Dynamic Pricing for Improvement of Ridesharing Platform Service in Seoul. Electronics. 2020; 9(11):1818. https://doi.org/10.3390/electronics9111818

Chicago/Turabian Style

Song, Jaein, Yun Ji Cho, Min Hee Kang, and Kee Yeon Hwang. 2020. "An Application of Reinforced Learning-Based Dynamic Pricing for Improvement of Ridesharing Platform Service in Seoul" Electronics 9, no. 11: 1818. https://doi.org/10.3390/electronics9111818

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop