Congestion Pricing - Links to Congestion Pricing Home

3.0 NATIONAL EVALUATION OVERVIEW

This chapter summarizes how the national evaluation of the UPA sites is being organized and carried out and identifies the steps in the San Francisco UPA evaluation process.

3.1 National Evaluation Organizational Structure

The evaluation of the UPA/CRD national evaluation is sponsored by the U.S. DOT. The Research and Innovative Technology Administration's (RITA) ITS JPO is responsible for the overall conduct of the national evaluation. Representatives from the modal agencies are actively involved in the national evaluation.

Members of the Battelle evaluation team include:

  • Battelle Memorial Institute – Prime;
  • Texas Transportation Institute (TTI), The Texas A&M University System;
  • Center for Urban Transportation Research (CUTR), University of South Florida;
  • Hubert H. Humphrey Institute of Public Policy and Center for Transportation Studies (CTS), University of Minnesota;
  • Eric Schreffler, ESTC; and
  • Susan Shaheen and Caroline Rodier, University of California, Berkeley.

As highlighted in Figure 3-1, the Battelle team is organized around the individual UPA/CRD sites. A site leader is assigned to each site, along with specific Battelle team members. The site teams are also able to draw on the resources of 4T experts and evaluation specialists.

The purpose of the national evaluation is to assess the impacts of the UPA/CRD projects in a comprehensive and systematic manner across all sites. The national evaluation will generate information and produce technology transfer materials to support deployment of the strategies in other metropolitan areas. The national evaluation will also generate findings for use in future federal policy and program development related to mobility, congestion, and facility pricing.

The focus of the national evaluation is on assessing the congestion reduction realized from the 4T strategies and the associated impacts and contributions of each strategy. The non-technical success factors, including outreach, political and community support, institutional arrangements, and technology will also be documented. Finally, the overall cost benefit analysis of the deployed projects will be examined.

Members of the Battelle team are working with representatives from the local partner agencies and the U.S. DOT on all aspects of the national evaluation. This team approach includes the participation of local representatives throughout the process and the use of site visits, workshops, conference calls, and e-mails to ensure ongoing communication and coordination. The local agencies are responsible for data collection, including conducting surveys and interviews. The Battelle team is responsible for providing the local partners direction on the needed data, formats and collection methods and for analyzing resulting data and reporting results.

Figure 3-1. Battelle Team Organizational Structure. A chart shows lines of communication linking the U.S. DOT UPA Evaluation Team to the COTM, down to Project Manager, and down to Principal Investigator and Deputy PM. From here, the line extends in two branches, Site Leaders and Site Specific Evaluation Teams on the left, and National Evaluation 4T Experts on the right.

Figure 3-1. Battelle Team Organizational Structure

3.2 National Evaluation Process and Framework

The Battelle team developed a National Evaluation Framework (NEF) to provide a foundation for evaluation of the UPA/CRD sites. The NEF is based on the 4Ts congestion reduction strategies and the questions that the U.S. DOT seeks to answer through the evaluation. The NEF is essential because it defines the questions, analyses, measures of effectiveness, and associated data collection for the entire UPA/CRD evaluation. As illustrated in Figure 3-2, the framework is a key driver of the site-specific evaluation plans and test plans and will serve as a touchstone throughout the project to ensure that national evaluation objectives are being supported through the site-specific activities.

Figure 3-2. The National Evaluation Framework in Relation to Other Evaluation Activities. A flow chart shows movement of information upward across five levels. At the bottom is the National Evaluation Framework. The next level is Site-Specific Evaluation Plans and Test Plans. in parallel with Review Evaluation Plans. The middle level includes Data Collection and Analysis associated with the State of Minnesota and with the cities of San Francisco, Seattle, and Los Angeles, in parallel with Miami Monitor and Support. The next level include Evaluation Reports for the cities of San Francisco, Seattle, and Los Angeles, in parallel with Miami Evaluation Report. The top level is the National Evaluation Findings Report.

Figure 3-2. The National Evaluation Framework in Relation to Other Evaluation Activities

The evaluation of each UPA/CRD site will involve several steps. With the exception of Miami, where the national evaluation team is serving in a limited role of review and support to the local partners, the national evaluation team will work closely with the local partners to perform the following activities and provide the following products:

  • a site-specific strategy guided by the NEF;
  • a site-specific evaluation plan that describes the strategy and provides a high-level view of all the test plans needed, the roles and responsibilities, and the schedule;
  • multiple site-specific test plans that provide complete details on how the data collection and analysis activity will be implemented;
  • collection of one year of pre-deployment and one year of post-deployment data;5
  • analysis of the collected data; and
  • site-specific evaluation reports and a National Evaluation Findings Report.

The NEF provides guidance to the local sites in designing and deploying their projects, such as by identifying the need to build in data collection mechanisms if such infrastructure does not already exist. To measure the impact of the congestion strategies, it is essential to collect both the “before” and “after” data for many of the measures of effectiveness identified in the NEF. Also important is establishing as many common measures as possible that can be used at all of the sites to enable comparison of findings across the sites. For example, a core set of standardized questions and response categories for traveler surveys will be prepared. Questions may need to be tailored or added to reflect the specific congestion strategies and local context for each site, such as road names or transit lines, but striving for comparability among sites will be a goal of the evaluation.

A traditional “before and after” study is the recommended analysis approach for quantifying the extent to which the strategies affect congestion in the UPA/CRD sites. In the “before,” or baseline condition, data for measures of effectiveness will be collected before the deployments become operational. For the “after” or post-deployment period, the same data will be collected to examine the effects of the strategies. The analysis approach will track how the performance measures changed over time (trend analysis) and examine the degree to which they changed between the “before” and “after” periods. Whenever possible, field-measured data will be used to generate the measures of effectiveness.

3.3 U.S. DOT Four Questions and Mapping to 12 Analyses

Table 3-1 shows the four “Objective Questions” that U.S. DOT has directed the national evaluation team to address.6 The analyses present what must be studied to answer the four objective questions. Table 3-2 identifies the 12 evaluation analyses described in the National Evaluation Framework and shows how they relate to the four objective questions. These 12 analyses form the basis of the evaluation plans at the UPA/CRD sites, including San Francisco.

Table 3-1. U.S. DOT National Evaluation “Objective Questions”
# Question
Objective Question #1 How much was congestion reduced in the area impacted by the implementation of the tolling, transit, technology, and telecommuting strategies?? It is anticipated that congestion reduction could be measured by one of the following measures, and will vary by site and implementation strategy:
  • reductions in vehicle trips made during peak/congested periods;
  • reductions in travel times during peak/congested periods;
  • reductions in congestion delay during peak/congested periods; and
  • reductions in the duration of congested periods.
Objective Question #2 What are the associated impacts of implementing the congestion reduction strategies? It is anticipated that impacts will vary by site and that the following measures may be used:
  • increases in facility throughput during peak/congested periods;
  • increases in transit ridership during peak/congested periods;
  • modal shifts to transit and carpools/vanpools;
  • traveler behavior change (e.g., shifts in time of travel, mode, route, destination, or forgoing trips);
  • operational impacts on parallel systems/routes;
  • equity impacts;
  • environmental impacts;
  • impacts on goods movement; and
  • effects on businesses.
Objective Question #3 What are the non-technical success factors with respect to the impacts of outreach, political and community support, and institutional arrangements implemented to manage and guide the implementation?
Objective Question #4 What are the overall costs and benefits of the deployed set of strategies?

Table 3-2. U.S. DOT Objective Questions vs. Evaluation Analyses
U.S. DOT 4 Objective Questions Evaluation Analyses
#1 – How much was congestion reduced? #1 –?Congestion
#2 – What are the associated impacts of the congestion reduction strategies? Strategy Performance #2 –?Strategy Performance:? Tolling
#3 – Strategy Performance:? Transit
#4 –?Strategy Performance:? Telecommuting/TDM
#5 –?Strategy Performance:? Technology
Associated Impacts #2 –?Strategy Performance:? Tolling
#3 – Strategy Performance:? Transit
#4 –?Strategy Performance:? Telecommuting/TDM
#5 –?Strategy Performance:? Technology
#3 –?What are the non-technical success factors? #11 –?Non-Technical Success Factors
#4 –?What is the overall cost and benefit of the strategies? #12?–?Cost-Benefit Analysis

The analyses associated with Objective Question #2 are of two types. The first four analyses focus on the performance of the deployed strategies associated with each of the 4Ts. These analyses will examine the specific impacts of each deployed project/strategy, and, to the extent possible, associate the performance of specific strategies with any changes in congestion. The second type of analysis associated with Objective Question #2 focuses on specific types of impacts, e.g., “equity” and “environmental.”

The 12 evaluation analyses were further elaborated into one or more hypotheses for testing. In some cases, where the analysis is not guided by a hypothesis, per se, such as the analysis of the non-technical success factors, specific questions are stated rather than hypotheses. Next, measures of effectiveness (MOE s) were identified for each hypothesis, and then required data for each MOE .

3.4 San Francisco UPA National Evaluation Process

Figure 3-3 presents the San Francisco UPA national evaluation team. The team includes the Contract Officer Technical Manager (COTM) who serves as the U.S. DOT National Evaluation leader, the U.S. DOT evaluation team, the FHWA point of contact for the site, and the Battelle team. The national evaluation team works with representatives from the partnership agencies, shown previously in Section 2, in development of the UPA evaluation for San Francisco.

Figure 3-3. San Francisco UPA National Evaluation Team. A chart shows lines of communication linking the U.S. DOT UPA Evaluation Team to the COTM, down to Project Manager, and down to Principal Investigator and Deputy PM, and on to the Site Leader.  A separate line goes from the COTM through the San Francisco Evaluation Points of Contact to the Site Leader. The Site Leader also has a side link to the 4T Experts and the Evaluation Specialists, as well as a link down to the Site Evaluation Team.

Figure 3-3. San Francisco UPA National Evaluation Team

Figure 3-4 presents the process for developing and conducting the national evaluation of the San Francisco UPA projects. The major steps are briefly discussed following the figure.

Figure 3-4. San Francisco UPA National Evaluation Process. A vertical milestone chart indicates the timeline from Kick-off Conference Call on January 26, 2009 to the Analysis and Evaluation Reports from Spring 2011 to December 2011. Workshops, Meetings, Plans, and other activities are listed.

Figure 3-4. San Francisco UPA National Evaluation Process

Kick-Off Conference Call. The kick-off conference telephone call, held on January 26, 2009, introduced the San Francisco partners, the U.S. DOT representatives, and the Battelle team members. The San Francisco UPA projects and deployment schedule were discussed, and the national evaluation approach and activities were presented. A PowerPoint presentation and various handouts were distributed prior to the conference call.

Site Visit and Workshop. Members of the U.S. DOT evaluation team and the Battelle team convened with the San Francisco partners in the Bay Area on February 17 and 18. The first day was used by the partners to brief the evaluation team on the UPA projects. SFMTA provided a tour of selected SFpark zones. A day-long evaluation workshop was held on the second day. Members of the U.S. DOT, Battelle, and local agency teams discussed potential evaluation strategies, including analyses, hypotheses, data needs, and schedule. A PowerPoint presentation containing the preliminary evaluation strategy, analysis, data needs, and other information was distributed prior to the workshop. A summary of the workshop discussion was prepared and distributed to participants after the workshop.

San Francisco UPA National Evaluation Strategy. The San Francisco UPA national evaluation strategy was revised based on the discussion at the workshop and the completion of the National Evaluation Framework. The San Francisco UPA evaluation strategy included the hypotheses/ questions, measures of effectiveness, and data needs for the analysis areas. The strategy also included a preliminary pre- and post-deployment data collection schedule, possible issues associated with the evaluation, and approaches for addressing exogenous factors. The San Francisco UPA national evaluation strategy was presented in a PowerPoint presentation, which was distributed to representatives of the U.S. DOT team and the San Francisco partners and a conference call was held on April 24 to review and discuss the evaluation strategy. There was agreement among all parties on the San Francisco UPA evaluation strategy and formal approval from the U.S. DOT was subsequently received to proceed with development of the San Francisco UPA national evaluation plan.

San Francisco UPA National Evaluation Plan. This document constitutes the San Francisco UPA national evaluation plan. The report provides a background to the U.S. DOT UPA, describes the San Francisco UPA projects, and presents the San Francisco UPA evaluation plan and preliminary test plans. The report was distributed in late August 2009 and reviewed with U.S. DOT and San Francisco UPA partners during a conference call on October 17, and the final plan is based on all comments and discussions about the evaluation plan. The document will guide the overall conduct of the San Francisco UPA national evaluation.

San Francisco UPA National Evaluation Test Plans. Based on approval from the U.S. DOT, the Battelle San Francisco UPA evaluation team will proceed with developing separate, more detailed test plans for each type of data needed for the evaluation, i.e., traffic, parking, etc. The preliminary test plans contained in the evaluation plan provide the basis for the more fully-developed test plans. Between December 2009 and February 2010 the individual test plans will be developed and reviewed with representatives from the U.S. DOT and local partnership agencies.

Pre-Deployment Data Collection. Based on approval of the San Francisco UPA evaluation individual test plans, data collection activities for the pre-deployment period will be initiated. The schedule of deployment for SFpark necessitates an abbreviated pre-deployment data collection period. However, where archived data are available and helpful in establishing long-term trends and in assessing the influence of exogenous factors (such as gas prices), they will be utilized. As discussed in Section 2, the individual projects will come on-line at various points between April and December 2010. Thus, the pre-deployment timeframe will begin in January 2010 and end as early as April and as late as December 2010 depending upon the project.

Post-Deployment Data Collection. Collection of post-deployment data of the San Francisco UPA projects will begin when the SFpark zones to be evaluated become operational in 2010. As other UPA projects come on-line in 2010, such as dissemination of parking pricing information on 511 phone in April and the 511 website and TransLink® in garages in December, post-deployment data collection for those projects will begin. After the last project is deployed, in December 2010, the final post-deployment data collection will take place. Thus, the post-deployment data collection period stretches from April 2010 through the spring of 2011.

Analysis and Evaluation Reports. Analysis of baseline data will begin once all of the data have been collected in the spring of 2011. Analysis of early (e.g., the first several months of) post-deployment data will begin shortly after the beginning of post-deployment data collection in mid-2010. A technical memorandum on evaluation early results, based on four or five months of post-deployment data, will be completed in the fall of 2010. The final evaluation report will be completed by December 2011.


5 While one-year each of pre- and post-deployment data are desirable, the operational data for specific projects within the overall evaluation schedule may result in more or less than a year's data being collected either pre- or post-deployment.
6 "Urban Partnership Agreement Demonstration Evaluation - Statement of Work," United States Department of Transportation, Federal Highway Administration; November 29, 2007.