Office of Operations
21st Century Operations Using 21st Century Technologies

Multimodal System Performance Measures Research and Application: Innovation and Research Plan

Chapter 3. Research Findings

Literature Review

The literature review found no "ideal" or universal multimodal transportation system performance measure or approach. There is clearly a strong focus on measuring performance through speed, delay, and reliability. The identified measurement processes focus most substantially on the performance of the highway system, for reasons of historical priority naturally buttressed by the most prolific data, now being enhanced through crowdsourced data from GPS devices. Transit performance measurement is making progress, aided by the increased use of automated vehicle locator (AVL) technology that provides continuous information on vehicle position that can be used to determine speed, delay, and on-time performance. The same types of technology innovations have not made their way to the non-motorized modes, although automatic counting devices are beginning to return data on usage for some bicycle facilities.

Discussions with a wide cross-section of experts and practitioners in the field supported the findings of the literature review: no method truly compares and accounts for all of our modal options in the planning and programming process, which sustains modal thinking and program silos. At the same time, discussions indicated a high level of interests in pursuing, defining and testing such a method.

An "Ideal" Measure

The literature review task confirmed the findings of the project work statement that transportation system performance measurement continues to be modally oriented. The orientation is primarily due to data limitations, resulting in a mode specific, facility-based analytical framework. This framework has become so ensconced in the profession that it is taken as a given. Thus, an "ideal" measure faces not only data challenges but conceptual and acceptance challenges as well.

Chapter 2 presents the reasons why a complete person trip perspective is best suited for an "ideal" multimodal measure. Complete person trips reflect a traveler's perspective, provide a rational basis for defining the system, and reduce aggregation challenges. The "ideal" measure should also be easily understood by professionals and travelers and provide feedback on system performance at any selected location, regardless of size, across the United States. Finally, because it provides a new perspective, the "ideal" measure can become a capstone for a new set of related, complete, trip-based measures.

The research team developed MSP as the "ideal" measure of multimodal transportation system performance. Productivity is the ratio of inputs to outputs in a production process and, for transportation systems, completed person trips are outputs, and network travel times are production inputs. The MSP is the number of completed person trips per minutes over a selected network/area and during a selected travel time. The higher the score, the higher the productivity.

Table 1 illustrates a calculation method for the MSP (others are certainly possible). Eight differing complete person trips are listed in table 1 along with travel times for each under three scenarios. Trips per minute for each trip is totaled across all trips to calculate the MSP. The first scenario presents a base condition. In scenario 2, the travel times for trips 4 and 5 increase, reducing the MSP to 0.26. In scenario 3, scenario 1 travel times stay the same, but a walk trip is added to raise the MSP to 0.32. It is worth noting that even though the added walk trip is slow, its overall travel time is consistent with the other trips, thereby raising the score.

Table 1. Multimodal system productivity calculation.
Person Trip
Travel Mode
Scenario 1
Scenario 2
Scenario 3
Minutes
Trips/Minute
Minutes
Trips/Minute
Minutes
Trips/Minute
1
Auto
40
0.03
40
0.03
40
0.03
2
Auto
20
0.05
20
0.05
20
0.05
3
Auto
35
0.03
35
0.03
35
0.03
4
Auto
30
0.03
40
0.03
30
0.03
5
Auto
25
0.04
40
0.03
25
0.04
6
Auto
20
0.05
20
0.05
20
0.05
7
Auto
45
0.02
45
0.02
45
0.02
8
Auto
30
0.03
30
0.03
30
0.03
0
Walk
NA
NA
NA
NA
30
0.03
MSP
No Value
245
0.28
300
0.26
275
0.32

The MSP measure is designed to provide feedback on the three performance perspectives listed in the last chapter:

  • System productivity – the MSP is a direct measure of system productivity and table 1 illustrates how it quantifies productivity.
  • Network efficiency – in concept, efficiency would be measured by dividing the MSP by network miles. To illustrate the concept, assume that over a year the 80th percentile MSP score for area A during the 8AM to 9AM hour is 10,000, and that the total number of network miles (roadway lanes, transit routes, bicycle and pedestrian paths) during that same time period is 100. The resulting efficiency ratio is 1,000 completed person trips per minute per mile. In area B, the 80th percentile MSP score for the same time period is 18,000 over a 200-mile multimodal network, resulting in an efficiency ratio of 900 completed person trips per minute per mile, slightly lower than area A.
  • System resiliency – in concept, resiliency would be measured by comparing MSP variations over time. For example, in area A during the mid-week AM peak period over the course of a year the average MSP is 8,000 person trips per minute with a standard deviation of 800. In area B over the same time period, the MSP average is also 8,000, but the standard deviation is 1,600, reflecting a higher frequency of disruptions, or lower network resiliency, or both in area B than in area A.

Data Gap Analysis

As noted in chapter 2, the MSP requires continuous and traced complete trip data. It also requires detailed network information. Standardized and detailed network data are readily available from several propriety and non-proprietary sources. Non-proprietary sources include Open Street for roadways, bicycle, and pedestrian facilities and General Transit Feed Specification (GTFS) for transit.

Complete person trip information is more elusive. Private companies such as Google, Apple, and Cubq, are collecting the data from mobile phones and GPS devices. GPS data provides accurate information but with small sample sizes. Based on GPS data, location-based services (LBS) data from mobile phone apps provide the same accuracy for an increasing number of trips.
Companies such as Streetlytics, StreetLight, and Mobility Labs, are expanding GPS and LBS data into complete trip information using algorithms that weight sampled trips using points of interest data from the U.S. Census Bureau and private sources and other data. At the time of the research, only hourly auto vehicle trips averaged over a month were available from those vendors. It was speculated and confirmed by conversations with data vendors that Google collects large samples of complete person trips not only across the United States but across the world. Sample sizes would likely provide the type of continuous, multimodal data needed by the MSP, but privacy and cost issues make obtaining such data challenging.

The recommended MSP measure requires continuous and traceable complete trip information. Given the current limitations of such data, the research shifted to assessing the extent and quality of more conventional travel data collected through electronic traffic counting technology, transit AVL and passenger counter (known as APC) systems and GTFS-fed traveler information systems, and bicycle-pedestrian counting programs. Those data require surrogate measures, ones that attempt to replicate the MSP with the limited data. The first surrogate measure is total person trips, which approximates the MSP by estimating total system demand for any given time period. Person trips is the summation of total trips across study area segments. The second is "weighted person trips," which approximates the productivity of travel across the network. Weighted person trips is estimated in three steps, the first calculating PMT by segment (segment person trips multiplied by segment length), the second summing PMT across facility segments within the pilot test study area, and the third dividing the summed PMT by the total travel time across those same facility segments.

Pilot Tests

The three pilot tests conducted for this research determined the quality of data collected and the feedback from surrogate measures described in the last section. Specifically, the pilot tests determined how well the surrogate measures answered questions such as:

  • How complete and accurate are collected data?
  • How effective is the feedback from the measures in determining how well the system operates in concert to serve travelers, particularly during peak and disrupted periods?
  • Can the measures provide insights on the source of problems and potential solutions?

Pilot Test Locations

A national scan of potential pilot test sites proved challenging because of the lack of multimodal data at any potential location. The lack of data was particularly true for bicycle and pedestrian data. The scan resulted in three pilot locations:

  • The San Marcos to Escondido corridor north of San Diego (Figure 4) – This pilot study measured performance along a 5-mile multimodal corridor, which includes an expressway, arterials, light rail transit line, bus routes, and a dedicated bicycle and pedestrian trail.
  • Crystal City in Arlington, VA (Figure 5) – This pilot measured the performance of travel into and out of a high-intensity, mixed-use center on the edge of downtown Washington, D.C. The area includes a robust multimodal network including a major arterial (US 1), local streets, Metro subway, Arlington bus routes, and bicycle and pedestrian connections.
  • Downtown Philadelphia (Figure 6) – Like Crystal City, this pilot measured multimodal travel into and out of the Philadelphia Central Business District (CBD).

The pilot tests measured performance from differing perspectives. The San Diego pilot tested the surrogate measures across a multimodal corridor. In both the Crystal City and Philadelphia pilots, the perspective changed to measuring performance of trips beginning or ending in the study area.

The San Diego corridor was divided into three segments. The California Department of Transportation (Caltrans) provided the SR 78 count and travel time data, the North County Transit District (NCTD) provided count and travel-time data for its Sprinter rail and local bus transit systems. Bicycle and pedestrian counts, not travel times, were provided by the San Diego Association of Governments (SANDAG) for three locations along the corridor's bicycle and pedestrian trail.

In Philadelphia, the Delaware Valley Regional Planning Commission (DVRPC) collected highway, transit, bicycle, and pedestrian count data along a cordoned area surrounding downtown Philadelphia. The count locations have remained consistent during differing data collection periods so that DVRPC could track changes in travel among modes over time. Data collection periods extended over short periods (1 or 2 days at most). DVRPC did not collect travel times.

The study team developed roadway data in Crystal City from crowdsourced data provided by StreetLight, which were checked against a limited set of traffic count data provided by Arlington County. The StreetLight data was averaged by hour and month, and included segment vehicle counts, segment travel times, and origin-to-destination travel times and distances. The Washington Metropolitan Area Transit Authority (WMATA) provided tap-on and tap-off records for riders either entering or exiting the Metro transit system at the Pentagon City and Crystal City stations. Bus data from both WMATA and Arlington Transit provided daily on and off information for study area stops, but not information about travel times on the system. Arlington County provided bicycle and pedestrian count information from four locations in the study area. Bicycle and pedestrian travel time data were not available.

A map of the area from Escondido to San Marcos (California).  Colored lines denote bus transit, SR 78, Mission Road and a bike-ped trail.  The map is described in surrounding text.
Figure 4. Map. Escondido to San Marcos (California) study corridor.3

A map of Crystal City (Arlington, VA).  The map denotes the Metro station and metro lines, roadway network, bus routes, and various types of bike trails and routes.
Figure 5. Map. Crystal City (Arlington, VA) study area.4

Map of downtown Philadelphia study area with border area denoted.  The map is discussed in the surrounding area. Figure 6. Map. Downtown Philadelphia study area (image provided by Delaware Valley Regional Planning Commission).5

Results

Data. To the extent they could be cross-checked, the facility-based, pilot-test data appeared accurate. The biggest data issues were inconsistencies in the timing, location, and types of travel data, making it difficult to paint a complete picture of multimodal performance for any of the pilots.

Measures. Despite inconsistent and incomplete data, the surrogate measures provided both expected and insightful feedback on multimodal performance. Insights of note include the following:

  • Both surrogate measures accurately reflected roadway density and flow dynamics. Results during congested conditions on major roadways align with density and flow theories; vehicle densities reached a breaking point where both density (total trips) and flow (trips per minute) drop.
  • Results from both surrogate measures suggested that vehicle trips divert from major roadways and use alternative travel paths during peak periods. Results indicated the larger roadway networks performed better than the major roadway in the network, suggesting that the typical practice of focusing on the performance of a major facility may overstate the loss of system productivity. Alternative routes appear to provide additional productivity that offsets, to differing degrees, the loss of productivity on the major road.
  • Results from both surrogates indicate a performance boost from dedicated guideway transit during peak periods. Guideway transit accommodated a higher proportion of person trips during peak demand and did so with greater productivity.
  • Bus transit and bicycle and pedestrian trips exhibit the same weekday peaking patterns as the other modes, but for the San Diego and Crystal City pilot sites tested, those modes did not add significantly to total person trips (less than 3 percent). Walking and bicycle trips (5 and 1 percent of total trips, respectively) provided a greater increase in person trips in Philadelphia, but were still comparatively low.
  • Deviations in total person trips and weighted person trips were lower than anticipated. Major road deviations were highest during peak hours, likely due to saturation dynamics, while major transit deviations were lowest during peak hours, likely due to the consistency of commuting trips.
  • While the measures did reflect extended disruptions, such as holidays and snow days, they did not demonstrate travel path and/or mode shifts during short-lived disruptions. This could be due to the coarseness of the data. It is also possible the measures accurately reflect the lack of travel path and modal shifts due to relatively short-lived disruptions. Travelers are committed to travel modes, and possibly paths, when incidents occur. It is possible that both are true.

Observations

Completeness and Quality of Data. The notable data problem uncovered by the pilot tests was the challenge of obtaining complete trip data. At the time of the tests, only complete auto vehicle trip data were available, and those data were provided as hourly averages by month, not continuous over small time slices. Because complete trip data were not available, pilot test results were unable to test the MSP measure (completed person trips per network minute).

The pilot tests uncovered other data problems, most notably the lack of coordination among agencies and modes when collecting travel data. An exhaustive national scan of potential pilot test sites identified only one location, downtown Philadelphia, with a coordinated multimodal data collection effort. The scan made it apparent that bicycle and pedestrian data are scarce, thus the availability of those data became the primary criterion when contacting agencies. Even with reasonable bicycle and pedestrian data, it was difficult to find locations where data from differing modes were collected during the same time periods, indicating that most data collection efforts support single mode performance measurement, not multimodal measurements.

System Definition

Confirming the system definition challenge highlighted in chapter 2, defining the pilot test networks was difficult because of the lack of information about travel demand and how it related to the selected networks. The study team defined the networks for two of the pilot studies (the Philadelphia pilot area was defined by the DVRPC). The San Marcos to Escondido study area was easier to define because it was larger than Crystal City (demand and network differences increase as the geographic focus increases) and because the network has a one-dimensional (east- west) orientation. Experience with travel demand in similar corridors, rather than a data based rationale, played a part in how to bound the study area. Crystal City presented a greater challenge because it was smaller and was designed to reflect how well the transportation system serves a place-based destination. Decisions about how to define the network were complicated. The most clear-cut decision was to include only ramps and not the mainlines of the expressways on the edge of Crystal City. This eliminated through trips from the analysis. Such decisions could not be made for arterials and transit.

When reporting on the relationships for major roads and premium transit in both the San Marcos to Escondido corridor and Crystal City, it had to be assumed that those networks served the same travel markets, a necessary and often taken-for-granted assumption in segment-oriented performance measures. It was possible in Escondido and Crystal City that the roadway and transit travel markets differ significantly, thereby leading to the erroneous conclusion that the major facilities in both areas worked together to serve demand. This potential travel market disparity was particularly true in Crystal City, where the network extended in several directions. To address this issue, the Crystal City pilot added an assessment of travel along the north-south US-1 and Metrorail corridor. Even then, the relationships between travel markets on both facilities likely differed because of differences in the areas served by the roadway and transit networks.

Aggregation Challenges. For reasons described in chapter 2, aggregating modal segment information into a single system measure was difficult. For the person trip surrogate measure, screenlines were used to account for over-counting person trips on segments. For the weighted person trip measure, segment person trips were multiplied by segment lengths and divided by travel times across aggregated segments.

In addition, it was difficult to create apples to apples aggregations across modes. This was particularly true for Crystal City where the network extends in different directions. Ultimately, the Crystal City segments were organized around the US-1 and Metrorail corridor.

What If?

Assuming complete person trip data were available for all three sites, it is likely the pilot test analysis could have been simpler and the feedback more straightforward and understandable. Areas, not networks, would have been selected and subdivided by parcels, geography based on U.S. Census Bureau data, or traffic analysis zones. Complete person trip data for a defined time period would have been collected for each subdivision with dynamic networks defined by the traced paths of those trips. An MSP score for each subdivision would have been determined by first calculating the number of person trips traveling between each subdivision (origin) and destination (origin-destination, or OD, pair), then dividing the total by the travel time between the OD pair, then summing the scores for all pairs (similar in concept to the example in table 1). The areawide MSP would be the summation of all OD pair scores. The overall MSP score could have been stratified in many ways, such as by mode, corridor, destination type, etc. The MSP scores could have provided insightful feedback on network efficiency and resiliency. While those place-based MSP scores could indicate poor performance to and from a given area, traced path information, in combination with facility-based performance analysis, would be needed to identify specific network problems causing the poor performance.

3 Baes map provided by Google, overlays by Renaissance Planning [ Return to Note 3 ]

4 Map created by Renaissance Planning [ Return to Note 4 ]

5 Map provided by Delaware Regional Planning Commission [ Return to Note 5 ]

Office of Operations