In sum, the contemporaneous record does not include any
information to support the conclusion that the agency, in
making its source selection decision, performed a
meaningful, qualitative assessment or critical,
comparative analysis of the proposals under the technical
evaluation factor or its enumerated elements.
Against this backdrop, the record shows that there were
substantial differences in the proposed staffing offered
by M7 and DS2. Specifically, the record shows that DS2
proposed an average of 22 percent fewer full time
equivalent (FTE) employees compared to the staffing
proposed by M7. This difference in proposed staffing is
approximately equal to the difference in the offerors’
respective prices; DS2’s price was approximately 21
percent lower than M7’s. Moreover, although the record
does not include any information about the other four
offerors, inasmuch as the record shows that M7 was
second-low offeror, it appears that DS2’s proposed
price‑and its proposed staffing--were substantially below
what the other four, higher-priced, offerors proposed.
In the final analysis, it may well be that the agency had
a reasonable basis for concluding, notwithstanding this
significant difference in the proposed staffing of DS2 and
M7, that the proposals nonetheless were technically equal.
However, in the absence of any explanation in the
contemporaneous evaluation record, we are left to guess at
the reasonableness of the agency’s conclusion. In
addition, and more fundamentally, the complete absence of
any critical analysis or qualitative assessment of the
proposals under the remaining elements of the technical
evaluation factor other than staffing also leaves us to
guess at the reasonableness of the agency’s broader
conclusion that all six proposals submitted were
technically equivalent under all of the RFP’s enumerated
technical evaluation elements. We therefore sustain M7’s
protest. (M7 Aerospace LLC
B-411986, B-411986.2: Dec 1, 2015) (pdf)
Swets
protests that the VA misevaluated the protester’s Lexi-Comp
and F&C databases, as presented in the product
demonstration. In particular, Swets contends that it
demonstrated that both its quoted products reached the RFQ-required
information in fewer screen selections than the VA’s
count. Swets also argues that the agency’s evaluation of
the number of screen selections required to reach the
required information was unreasonable, disparate from how
the agency treated Cox, and undocumented. Protest, July
11, 2014, at 30-33.
As explained below, we have been unable to determine from
this record how the agency’s evaluators reached the number
of screen selections for each of Swets’ databases at the
product demonstration. In this regard, the record contains
conflicting evidence, statements, and exhibits concerning
how the VA conducted its evaluation. As the agency failed
to maintain an evaluation record adequate to permit
meaningful review, and has failed to rebut persuasive
evidence presented by the protester about the number of
screen selections required to demonstrate the
effectiveness of Swets’ alternative databases, we sustain
the protest. Although we recognize that this procurement
was conducted as a request for quotations among BOA
holders under FAR subpart 16.7, it is a fundamental
principle of government accountability that an agency be
able to produce a sufficient record to allow for a
meaningful review where its procurement actions are
challenged. See Resource Dimensions, LLC, B-404536, Feb.
24, 2011, 2011 CPD ¶ 50 at 6; e-LYNXX Corp., B-292761,
Dec. 3, 2003, 2003 CPD ¶ 219 at 8; Checchi & Co.
Consulting, Inc., B-285777, Oct. 10, 2000, 2001 CPD ¶ 132
at 6. An agency which fails to adequately document the
rationale for its source selection, bears the risk that
its determinations will be considered unsupported, and
that absent such support, our Office may lack a basis to
find that the agency had a reasonable basis for its
determinations. Tiger Enters., Inc., B‑293951, July 26,
2004, 2004 CPD ¶ 141 at 2. In reviewing an agency’s
procurement actions, we do not limit our review to
contemporaneous evidence but consider, as appropriate,
hearing testimony and the parties’ arguments and
explanations. See Southwestern Marine, Inc.; Am. Sys.
Eng’g Corp., B-265865.3, B-265865.4, Jan. 23, 1996, 96-1
CPD ¶ 56 at 10.
As detailed above, the solicitation established that the
VA would use a product demonstration as the primary means
for determining the technical merit of each vendor’s
offering(s), and that the agency’s evaluation would be
based on, among other things, the number of screen
selections required to reach the desired information. RFQ
at 22-24. When conducting the product demonstration, the
agency evaluators and vendor representatives were not in
the same physical location; rather, the demonstrations
were conducted using screen-sharing software (e.g., “GoToMeeting,”
or “WebJoin”). AR, Tab 13, Declaration of VA Demonstration
Facilitator, Oct. 1, 2014, at 1. A separate audio line was
also established so that the agency evaluators and vendors
could communicate during the demonstration. Id. For each
of the required product demonstration items, the agency
evaluators instructed the vendor representatives to start
from the “home screen” and navigate to the screen
containing the required information. Id.
A Swets representative demonstrated the ability of the
Lexi-Comp and F&C products to search for and reach the
information required for each of the demonstration items.
Protest, July 11, 2014, Declaration of Swets Government
Contracts Director, July 11, 2014, at 3. The Swets
representative states that, after navigating to the
required information, he orally confirmed to the VA
evaluators his count of the number of screen selections
required (he also stated that the VA evaluators did not
dispute his count). Id.
In contrast to Swets’ claim, the agency evaluators state
that there was no verbal counting of screen selections by
the Swets product representative. AR, Tab 12, Declaration
of VA Evaluator J.N., July 28, 2014. The VA and Swets
agree, however, that the agency evaluators did not state
their count of the number of screen selections that it
took in each instance to reach the required information.
Protest, July 11, 2014, Declaration of Swets Government
Contracts Director, July 11, 2014, at 3.
The VA evaluators kept worksheets of their product
demonstration observations, including the total count of
the number of screen selections for each of the 15 items
that were researched during the demonstration. In most
instances the agency evaluators agreed with each other
about the number of screen selections needed to reach the
desired information for each of the required items. AR,
Tab 9, Evaluator Worksheets, May 16, 2014, at 1-30.
However, for the Lexi-Comp database product, item “c” (“[s]how
the mechanism of action for montelukast and provide
references”), two of the evaluators counted six screens,
while the third evaluator counted four to six screens. Id.
at 11, 14, 17.
Swets asserts that the agency’s evaluation of Swets’
screen selections was inaccurate and unreasonably high.
The protester also asserts that the agency’s evaluation of
screen selections was unequal. In support of its
allegations, the protester provided a declaration from the
representative who conducted the product demonstration.
The vendor representative states “I have used the
screen/click counting methodology [the VA] claims to have
used to calculate the number of screen selections/clicks
required for both Lexi-Comp and F&C to access the
information required by the evaluation factors, and the
results of my calculations . . . establish that the screen
counts reported by the [agency] evaluators were inaccurate
and were higher than warranted.” Protest, Aug. 21, 2014,
Declaration of Swets Government Contracts Director, Aug.
21, 2014, at 1-2.
To document his attempt at replicating the agency’s
evaluation methodology, Swets provided “screen-by-screen,”
“click-by-click,” walkthroughs (sometimes called “screen
shots”) for each of the different products at issue. For
example, with regard to demonstration item “a” (“Compare
in table format the contraindications/adverse reactions of
warfarin and dabigatran”), Swets showed that four screen
selections were necessary for Lexi-Comp to reach the
required information, in comparison to the VA’s count of
seven screen selections. Id., exh. 3, Lexi-Comp
Walkthrough for Item “a”. Likewise, with regard to
demonstration item “j” (“Display in table format the
contraindications/adverse reactions of prasugrel and
ticagrelor”), Swets showed that Lexi-Comp and F&C each
required four screen selections in comparison to the VA’s
count of eight and six screen selections, respectively.
Id., exh. 6, Lexi-Comp Walkthrough for Item “j”; exh. 7,
F&C Walkthrough for Item “j”. Swets also provided
screen-by-screen walkthroughs for demonstration item “c”
for Lexi-Comp (three screen selections as compared to the
VA’s count of six screen selections), and for
demonstration item “l” for F&C (three screen selections as
compared to the VA’s count of six screen selections). Id.,
exh. 11, Lexi-Comp Walkthrough for Item “c”; exh. 12, F&C
Walkthrough for Item “l”. Swets declares that for all
demonstration items, its total Lexi-Comp screen selection
count is 52 as compared to the VA total screen selection
count of 74, and that its F&C screen selection count is 51
as compared to the VA’s count of 61. Id., exh. 1, Screen
Counts for Lexi-Comp and F&C (Swets Count vs. VA Count).
In response, the VA asserts that its evaluation was
reasonable and consistent with the solicitation, and that
its counting standards were consistently applied to all
product demonstrations. The agency argues that the
evaluators agreed to the methodology for counting screen
selections prior to conducting the product demonstrations.
In this regard, the agency states, in its response to the
protest, that the evaluators counted the number of screen
selections as each screen was being clicked on and counted
transitioning to the next screen, which included all
drop‑down menus and sub-menus. Contracting Officer’s
Statement, Aug. 11, 2014, at 4; AR, Tab 12, Declaration of
VA Evaluator J.N., July 28, 2014, at 1. Thus, the agency
asserts that the evaluators accurately counted vendors’
screen selections based on this established methodology.
On this record, we cannot find that the agency’s
documentation in regard to its screen selection counts is
sufficient to allow us to review the reasonableness of the
agency’s evaluation. The contemporaneous evaluation record
consists of the evaluator worksheets which provide only
the total screen selection counts. The documents
subsequently submitted by the agency provide some
additional information regarding the agency’s methodology,
but provide no contemporaneous documentation of how the
agency reached the screen selection count that it did in
each instance.
In addition, the agency’s statements during the course of
the protest have not been consistent about the methodology
used by the evaluators. In sum, the agency essentially
argues that it treated all vendors equally as to its
screen selection counting, but it has no record to support
the validity of its count, other than the worksheets
showing the total count. In contrast, Swets has provided
detailed, screen-by-screen walkthroughs of the Lexi-Comp’s
and F&C’s products in support of its assertion that the
agency’s screen selection counts are inaccurate. In light
of the contrary information provided by the protester, the
agency evaluator worksheets that document only evaluation
results are not sufficient to demonstrate the
reasonableness of the agency’s evaluation. In sum, we
cannot find the agency’s evaluation results to be
reasonable because the agency has not adequately
documented the record on which its screen count results
are based. (Swets
Information Services, B-410078: Oct 20, 2014) (pdf)
Both
protesters point out, and the agency concedes, that
handwritten evaluation sheets from the evaluation team
members, although compiled into a summary document, were
not themselves retained. VA E-Mail to GAO, Jan. 22, 2014.
M‑Pak asserts that “VA’s failure to create or retain this
testing documentation makes it impossible for GAO or any
other impartial reviewer to assess the merits of the VA’s
evaluation.” M-Pak Comments at 4; see Custom Pak Comments
at 2.
We disagree. Where an agency fails to document or retain
evaluation materials, it bears the risk that there may not
be adequate supporting rationale in the record for us to
conclude that the agency had a reasonable basis for its
source selection decision. Navistar Def., LLC; BAE Sys.,
Tactical Vehicle Sys. LP, B‑401865 et al., Dec. 14, 2009,
2009 CPD ¶ 258 at 13. The destruction of individual
evaluator documents, however, does not render an agency’s
evaluation unreasonable per se; rather, we will consider
the record adequate if the consensus documents and source
selection decision sufficiently document the agency’s
rationale for the evaluations. Joint Mgmt. and Tech. Servs.,
B-294229, B-294229.2, Sept. 22, 2004, 2004 CPD ¶ 208 at
3-4.
Here, the evaluation record adequately detailed the
strengths and weaknesses that formed the basis for the
agency’s evaluation ratings for each vendor and for the
source selection decision. See National Beef Packing Co.,
B-296534, Sept. 1, 2005, 2005 CPD ¶ 168 at 12 n.6. The
agency has provided statements from the evaluators that
indicate that, although the evaluators’ notes were not
retained, they were taken into consideration and were
compiled into the summary source selection decision. In
this regard, the record indicates that:
[the program manager] was on the IPT [Integrated Product
Team] and participated in the USPS testing. She
consolidated the evaluation sheets for insertion into
the SSD. A summary document was provided to the team
during and at the conclusion of the evaluation process.
At no time did any of the team state their ratings were
not included.
Supp. AR, attach. 5 (Program Manager’s Statement I); see
also attach. 8 (Program Manager’s Statement II).
Further, the source selection decision sets forth a
detailed summary of the testing results for both the USPS
and the VA testing. SSDD at 3-10. With regard to the USPS
testing of M-Pak’s samples, the source selection decision
indicated the following results:
FAIL
Tears apart easily due to weak adhesive
Lots of damage to bags from sorting belts, especially
the Large and X‑Large bags
Does not withstand rough transit
SSDD at 4. With regard to the VA testing of M-Pak’s
samples, the overall results were as follows:
FAIL – The Small and Medium do not contain an
anti-static strip. The Large and X-Large bags have weak
glue and are not tamper-resistant. The X-Large bag has a
lighter inner liner that allows you to read the contents
in the package.
SSDD at 4; see also SSDD at 12. Similarly, with regard to
the VA testing of Custom Pak’s samples, the overall
results were as follows:
FAIL. Bags are not tamper-resistant, contain poorly
constructed side seams and mailing labels do not stick.
The material of the bags is “crinkly” and may impact
barcode readability through the mail stream.
SSDD at 7, 13. Further, these summary conclusions in the
source selection document are followed by detailed,
comprehensive evaluator findings with regard to each of
the four sizes of sample bags submitted by M-Pak, Custom
Pak, Star Poly, and the other vendors under each of the
five evaluation categories. See, e.g., SSDD at 4-5
(M-Pak), 7-8 (Custom Pak), 9-10 (Star Poly).
In these circumstances, given the comprehensive nature of
the agency’s summary evaluation documents, there is an
adequate supporting rationale in the record for us to
conclude that the agency had a reasonable basis for
finding the protesters’ bags noncompliant with the
specification requirements. This protest ground is denied.
(Custom Pak, Inc.; M-Pak, Inc.,
B-409308, B-409308.2, B-409308.3, B-409308.4: Mar 4, 2014)
(pdf)
Clark/F-P also
challenges the State Department’s evaluation of Harbert’s
final revised proposal, where the agency increased
Harbert’s proposal ratings from good to excellent under
Areas 1, 3, 4, and 6. The protester complains that there
is no documentation in the record explaining how Harbert’s
revisions adequately addressed the weaknesses the TEP had
identified in Harbert’s initial proposal and why Harbert’s
revised proposal merited an excellent rating under these
areas. Supp. Protest at 4.
For example, in Area 3, cost management and value
engineering, the TEP initially assessed Harbert’s proposal
as having 2 exceptional strengths and 1 weakness, and
assessed Clark/F-P’s proposal as having 5 exceptional
strengths and 3 strengths. See AR, Tab 8, TEP Initial
Consensus Evaluation Report, at 15, 35-36. Based on these
assessments, both Harbert’s and Clark/F-P’s proposals were
rated as good under Area 3. Id. After receiving revised
proposals, the TEP increased Harbert’s rating to
excellent, without explanation, other than the statement
that the “TEP accepted that the Offeror corrected the
weakness and the ranking was changed to Excellent.” See
AR, Tab 9, TEP Final Consensus Evaluation Report, at 3. In
this regard, Clark/F-P notes that the TEP did not identify
any further strengths in Harbert’s proposal as a result of
the firm’s proposal revisions.
The State Department responds that Harbert provided
comprehensive explanations in its revised proposal that
corresponded to each of the identified weaknesses, and
that the TEP reasonably increased Harbert’s proposal
ratings to reflect that there were no weaknesses in its
revised proposal. Supp. AR at 11, 16.
As the State Department correctly notes, the evaluation of
proposals and assignment of adjectival ratings should
generally not be based upon a simple count of strengths
and weaknesses, but on a qualitative assessment of the
proposals consistent with the evaluation scheme. Supp. AR
at 10 (citing Command Mgmt. Servs., Inc., B-310261,
B-310261.2, Dec. 14, 2007, 2008 CPD ¶ 29 at 4). Adjectival
ratings are merely a guide for intelligent decisionmaking.
One Largo Metro LLC; Metroview Dev. Holdings, LLC; King
Farm Assocs., LLC, B-404896 et al., June 20, 2011, 2011
CPD ¶ 128 at 14. However, evaluators and selection
officials should reasonably consider the underlying bases
for ratings, including the advantages and disadvantages
associated with the specific content of competing
proposals, in a manner that is fair and equitable and
consistent with the terms of the solicitation. See MD
Helicopters, Inc.; AgustaWestland, Inc., B-298502 et al.,
Oct. 23, 2006, 2006 CPD ¶ 164 at 15. Indeed, the Federal
Acquisition Regulation (FAR) requires that agencies
sufficiently document their judgments, including
documenting the relative strengths, deficiencies,
significant weakness, and risks supporting their proposal
evaluations. See FAR §§ 4.801(b), 15.305(a), 15.308;
Century Envtl. Hygiene, Inc., B-279378, June 5, 1998, 98-1
CPD ¶ 164 at 4.
Here, the record does not provide the underlying bases for
the TEP’s decision to increase the ratings of Harbert’s
proposal to excellent after revised proposals. As noted
above, the TEP final consensus evaluation report merely
states that the “TEP accepted that the offeror corrected
the weakness and the ranking was changed to Excellent”
without providing additional explanation. AR, Tab 9, TEP
Final Consensus Evaluation Report, at 2-3. The SSA also
testified that he had seen no written explanation for why
the TEP increased Harbert’s proposal rating from good to
excellent in these areas, nor did he engage in discussion
on this matter. Tr. at 80, 81. Given the lack of
documentation and support for the increased ratings in
Harbert’s proposal, we find no basis to conclude that the
TEP reasonably raised Harbert’s rating based on the firm’s
revised proposal. (Clark/Foulger-Pratt
JV, B-406627, B-406627.2, Jul 23, 2012) (pdf)
TEG contends that DOS’s past performance evaluation of
BlueLaw, BSOM, Crucible, FedSys, and GCJS was
unreasonable. Specifically, TEG contends that the agency
did not determine whether the submitted past performance
references of these offerors were relevant and permitted
these offerors to provide less than the required number of
references.
As a general matter, the evaluation of an offeror’s past
performance, including the agency’s determination of the
relevance and scope of an offeror’s performance history to
be considered, is a matter within the discretion of the
contracting agency, and we will not substitute our
judgment for reasonably based past performance ratings.
MFM Lamey Group, LLC, B-402377, Mar. 25, 2010, 2010 CPD ¶
81 at 10; Yang Enters., Inc., Santa Barbara Applied
Research, Inc, B-294605.4 et al., April 1, 2005, 2005 CPD
¶ 65 at 5. However, we will question an agency’s
evaluation conclusions where they are unreasonable or
undocumented. Clean Harbors Envtl. Servs., Inc.,
B-296176.2, Dec. 9, 2005, 2005 CPD ¶ 222 at 3. The
critical question is whether the evaluation was conducted
fairly, reasonably, and in accordance with the
solicitation’s evaluation scheme, and whether it was based
on relevant information sufficient to make a reasonable
determination of the offeror’s past performance. DRS C3
Sys., LLC, B-310825, B-310825.2, Feb. 26, 2008, 2008 CPD ¶
103 at 22.
TEG argues that the agency failed to consider the
relevance of the prior contracts of the awardees in
accordance with the RFP and the Federal Acquisition
Regulation (FAR). An agency is required to consider,
determine and document the similarity and relevance of an
offeror’s past performance information as part of its past
performance evaluation. See FAR § 15.305(a)(2); Clean
Harbors Envtl. Servs., Inc., supra.
While DOS in its report on the protest states that it
determined that the past performance of the awardees was
relevant and gave examples of why this was the case, there
is no contemporaneous documentation evidencing that the
TEP meaningfully considered whether these offerors’
references represented “relevant corporate experience
providing criminal justice related support services and
associated support systems required under Section C.” See
RFP § M.9.3.1.a at 268. Instead, the contemporaneous
documentation reflects that the TEP merely presumed that
each offerors’ identified contracts were relevant. For
example, as to the relevance of BlueLaw’s past
performance, the evaluation documentation only states:
BlueLaw identified their relevant past performance and
was rated Exceptional by two, Excellent by six, and Good
by one of their Past Performance Surveys submitted by
Contracting Officers that have worked directly with
them.
AR, Tab 10b, Post Discussion Consensus Report, at 15. The
past performance evaluations for the other offerors
included essentially identical language regarding the
presumed relevance of their past performance, and showed
that the TEP past performance evaluation focused primarily
on the quality of the offeror’s past performance records.
As set forth above, the RFP here required the agency to
consider whether an offeror and its major subcontractors
demonstrated relevant corporate experience providing
criminal justice related support services and associated
support systems required under the work statement in
section C. A review of BlueLaw’s past performance is
illustrative of the problems in DOS’s evaluation. BlueLaw
identified five prior contracts to demonstrate relevant
experience as a prime contractor: (1) “INL Program and
Outreach Support;” (2) “Enabling Human Rights For All;”
(3) “Towards the Full Inclusion of People with
Disabilities: Examining the Accessibility of Overseas
Facilities and Programs Funded by the United States;” (4)
“Paraguay Threshold Country Plan--Formalization of
Economic Activities;” and (5) “Iraq Civil Civilian
Advisory Support –Justice & Law Enforcement Assistance.”
See AR, Tab 4, vol. 4, Past Performance Client References,
at 1. There is no comtemporaneous documentation in the
record that reasonably explains why this past performance
meets the evaluation standards required by the RFP. In
addition, as the protester has pointed out, it is not
readily apparent that the identified efforts meet this
requirement.
Based on the foregoing, we cannot conclude on this record
that the agency’s past performance evaluation of the
proposals was reasonable. Accordingly, we sustain the
protest on this basis. (The
Emergence Group, B-404844.5,B-404844.6, Sep 26, 2011)
(pdf)
ITT
challenges the agency’s cost and technical evaluations and
maintains that, because the agency’s evaluation of
proposals was not reasonable, its best value source
selection also is unreasonable. We have carefully
considered all of ITT’s assertions and sustain its protest
for the reasons discussed below.
ITT principally maintains that, although the agency made
substantial adjustments to the offerors’ proposals in
performing its most probable cost evaluation, the record
contains no explanation for why the adjustments were made.
The protester also challenges the agency’s technical
evaluation, maintaining that the Army either misevaluated
its proposal or treated the offerors disparately during
the technical evaluation.
As a general matter, in reviewing protests against
allegedly improper evaluations, our Office examines the
record to determine whether the agency’s conclusions are
reasonable and consistent with the terms of the
solicitation and applicable statutes and regulations. SOS
Int’l, Ltd., B-402558.3, B-402558.9, June 3, 2010, 2010
CPD ¶ 131 at 2. Additionally, when an agency evaluates
proposals for the award of a cost-reimbursement contract,
an offeror’s proposed estimated cost of contract
performance is not considered controlling since,
regardless of the costs proposed by the offeror, the
government is bound to pay the contractor its actual and
allowable costs. Magellan Health Servs., B-298912, Jan. 5,
2007, 2007 CPD ¶ 81 at 13; Metro Machine Corp., B-295744,
B-295744.2, Apr. 21, 2005, 2005 CPD ¶ 112 at 9; Federal
Acquisition Regulation (FAR) § 16.301. Consequently, a
cost realism analysis must be performed by the agency to
determine the extent to which an offeror’s proposed costs
represent what the contract costs are likely to be under
the offeror’s unique technical approach, assuming
reasonable economy and efficiency. FAR §§ 15.305(a)(1),
15.404-1(d)(1), (2); The Futures Group Int’l, B-281274.2,
Mar. 3, 1999, 2000 CPD ¶ 147 at 3.
Finally, in considering an agency’s evaluation of cost and
technical proposals in a cost reimbursement setting, it is
axiomatic that the agency’s evaluation must be adequately
documented, such that the record reflects the agency’s
reconciliation of its technical and cost evaluation
conclusions. Serco, Inc., B-298266, Aug. 9, 2006, 2006 CPD
¶ 120 at 7. Based on our review of the record, we find the
agency’s evaluation unreasonable.
The record shows that, in performing its technical
evaluation, the agency identified [deleted] significant
strengths, [deleted] weaknesses, and no significant
weaknesses or deficiencies in the ITT proposal. AR, exh.
5, at 22-25. For Lockheed, the agency’s technical
evaluation identified [deleted] significant strengths,
[deleted] strengths, [deleted] weaknesses, and [deleted]
significant weaknesses or deficiencies. Id. at 25-28.
Among the overall total of [deleted] weaknesses identified
in both proposals combined, [deleted] were under the
staffing approach subfactor.
In performing its most probable cost evaluation, the
record shows that, for both proposals, the agency’s cost
adjustments were confined to changes in the offerors’
proposed staffing approach, and that the agency took no
exception to any other cost element proposed by either
offeror. AR, exh. 6, at 14-19, 41-45.
Both firms proposed [deleted] quantities of staffing to
perform the requirement, with ITT proposing [deleted]
staff hours and Lockheed proposing [deleted] staff hours.
Protester’s Comments, Nov. 17, 2011, at exh. A, attach. 3,
at 11. The record further shows that the agency made
significant changes to the offerors’ proposed staffing
when performing its most probable cost evaluation.
With respect to ITT, the agency increased its proposed
staffing in various areas by a total of [deleted] staff
hours and reduced its staffing in other areas by [deleted]
staff hours, for a combined change in ITT’s proposed
staffing of [deleted] staff hours. Protester’s Comments,
Nov. 17, 2011, at exh. A, attach. 1. These changes were
made in approximately [deleted] percent of the labor
categories ([deleted]) considered by the agency in its
evaluation of ITT’s proposal.
In evaluating the Lockheed proposal, the agency made
upward adjustments to its proposed staffing of [deleted]
staff hours and downward adjustments of [deleted] staff
hours, for a combined change in Lockheed’s proposed
staffing of [deleted]. Protester’s Comments, Nov. 17,
2011, at exh. A, attach. 2. These changes were made in
approximately [deleted] percent of the labor categories
([deleted]) considered by the agency in its evaluation of
the Lockheed proposal. Id.
Despite these significant changes in the offerors’
proposed staffing for cost evaluation purposes, the record
contains no explanation of why the changes were made. In
particular, the changes are memorialized in Excel
spreadsheets that reflect the agency’s mathematical
calculations of the changes, but these spreadsheets
include no narrative explanation of the reasons for the
changes. AR, exhs. 16a-16e. Additionally, although the
record includes a separate narrative cost report, the
report makes no mention of the reasons underlying the
agency’s staffing adjustments for cost evaluation
purposes, and instead simply describes the agency’s
approach to making the changes. For example, in the case
of the Lockheed proposal, the cost report simply states,
without elaboration, that:
[deleted]
AR, exh. 6, at 16. The record includes [deleted] language
describing the changes made to the ITT proposal. Id. at
43.
In addition to the lack of any explanation for the
extensive staffing changes in the agency’s cost evaluation
materials, an examination of the agency’s technical
evaluation report reveals no apparent connection between
the agency’s technical evaluation findings on the one
hand, and its cost evaluation findings on the other.
For example, the agency specifically identified [deleted]
weaknesses in its technical evaluation of the ITT proposal
relating to the firm’s use of [deleted] positions, finding
that these positions (which appear to total [deleted] full
time equivalents (FTEs)) are not required by the statement
of work (SOW). AR, exh. 5, at 23. Notwithstanding this
finding, the record shows that, in its cost evaluation,
the agency did not reduce ITT’s staffing for [deleted]
personnel. Protester’s Comments, Nov. 17, 2011, at exh. A,
attach. 1, at 1. Thus, the agency’s cost evaluation
appears inconsistent with the technical evaluation finding
that these positions are not necessary to perform the SOW.
In contrast, there were areas where, in its cost
evaluation, the agency concluded that Lockheed’s staffing
was significantly overstated (for example, the record
shows that the agency adjusted Lockheed’s proposed
[deleted] personnel downward by approximately [deleted]
FTEs, Protester’s Comments, Nov. 17, 2011, at exh. A,
attach. 2, at 2), but the agency did not assign the
proposal a weakness in its technical evaluation for this
identified overstaffing, as it did in evaluating the ITT
proposal in the area of computer operators. This
essentially amounts to disparate treatment of the two
firms, since the agency assigned a technical evaluation
weakness for overstaffing to one firm’s proposal, but not
to the other firm’s proposal.
In other instances, the agency made significant upward
adjustments to the offerors’ proposed staffing in numerous
areas, but the technical evaluation is silent with respect
to the apparent weaknesses that these staffing shortages
would suggest. For example, the agency increased
Lockheed’s proposed staffing for [deleted] by [deleted]
FTEs and its [deleted] personnel by [deleted] FTEs,
Protester’s Comments, Nov. 17, 2011, at exh. A, attach. 2,
at 2. Despite these significant upward adjustments to
Lockheed’s staffing for cost evaluation purposes, there is
no assignment of a weakness to the Lockheed proposal (or
even a mention in the agency’s technical evaluation
report) based on these significant staffing shortfalls.
In sum, the record shows that the agency made significant
changes to the offerors’ proposed staffing in the
overwhelming majority of their proposed labor categories
when conducting its most probable cost evaluation, but the
underlying basis for these changes is not memorialized in
the record. In addition, the record shows that the agency
failed to reconcile its evaluation in a manner that
demonstrates consistency between its cost and technical
evaluation findings. We conclude that the agency’s
evaluation of proposals is unsupported in the record and,
thus, is unreasonable. Correspondingly, the agency’s
source selection decision, based on this fundamentally
flawed evaluation, is likewise unreasonable. We therefore
sustain ITT’s protest. (ITT
Systems Corporation, B-405865,B-405865.2, Jan 6, 2011)
(pdf)
Past
Performance Confidence
For the past performance confidence factor, DISA concluded
that Solers' proposal merited a satisfactory confidence
rating, and BAH merited a substantial confidence rating.
The agency states that for this factor, the agency
considered the reference interviews, PPIRS data, and the
personal knowledge of the evaluators.
With regard to the evaluation of the past performance
references identified in the offerors' proposals, DISA
concedes that it cannot produce copies of the
questionnaires used during the interviews of the
references, nor does it have any record of the interviews
conducted with the references. SAR at 28; AR, Tab L, Cover
Memorandum, at 1; Email from DISA to GAO and Protester
(Jan. 27, 2011). Instead, the agency contends that the
evaluation summary in the SRD reflects the information
provided by the references to the evaluators. SAR at
27-28.
The SRD, however, merely lists conclusions by the
evaluators concerning each offeror's past performance. See
AR, Tab L, SRD, at 13-14. The record does not meaningfully
document information provided by the past performance
references, such as the questions asked and answers
received, or what strengths or weaknesses were associated
with particular past performance references. In the
absence of this information, we are unable to determine
whether the evaluation conclusions made by the evaluators
are supported by the information provided by the
references.
With regard to the PPIRS records, the record is unclear as
to how DISA used this data. For Solers, the agency
reviewed the PPIRS record for the incumbent CDWSG task
order, and records for two contracts other than those
identified by Solers in its proposal. Id.; DISA Response
to GAO Questions at 9. For BAH, the agency reviewed PPIRS
records for three contracts other than those identified by
BAH in its proposal. AR, Tab L, SRD, at 13-14.
In its response to the protest, DISA states that the PPIRS
data were used "to validate the information that was
provided in the interviews and to also assist in the
assignment of the overall relevance and confidence
ratings." DISA Response to GAO Questions at 9. In a
contemporaneous addendum to the SRD, the record includes a
number of statements prepared in connection with the
initial award evaluation and signed by the PPT members,
which state, "I conclude that our past performance
evaluation about [Solers or BAH] matches with the [PPIRS]
assessment." AR, Tab L, SRD Addendum, attach. 1, PPT
Members' Statements.
To the extent that the agency states that the PPIRS data
was used for the purpose of the substantive evaluation of
the offerors' past performance, the record provides no
support for this proposition. The SRD does not discuss how
the PPIRS information was used for Solers; instead, the
substantive conclusions regarding the protester's
performance relate solely to the undocumented interviews.
See AR, Tab L, SRD, at 13. For BAH, the SRD states that "[t]he
questionnaire and phone interview has identified that BAH
has excellent past performance," and that "[t]he PPIRS
data validate these findings." Id. at 14. Moreover, the
record does not document which PPIRS records or what
aspects of the contract performance records were
considered by the evaluators or found to support and/or
validate the ratings.
With regard to the evaluators' personal knowledge, DISA
contends that although the agency received information
concerning only one of BAH's references, the PPT utilized
the personal knowledge of its members concerning the other
two references. SAR at 30; DISA Response to GAO Questions
at 8-9. The agency concedes, however, that the
contemporaneous record does not discuss or otherwise
document the personal knowledge used by evaluators in
assessing the offerors' past performance. DISA Response to
GAO Questions at 9.
Past Performance Relevancy
With regard to the past performance relevancy ratings,
DISA states that its evaluation of the offerors considered
the reference interviews, the personal knowledge of the
evaluators, and the offerors' technical proposals. The
agency found that Solers' past performance was "relevant"
based on the following assessment:
The past performance citations provided by SOLERS
involved much of the magnitude of effort and
complexities this solicitation required. Certain
projects are not related cross domain [sic] such as [NCES]
Messaging and ASTG and hence the lower rating.
AR, Tab L, SRD, at 13. For BAH, the agency found that the
awardee's past performance was "very relevant" based on
the following assessment:
The past performance citations provided by BAH, were for
efforts that involved essentially the same magnitude of
effort and complexities this solicitation requires and
hence relevancy rating of Very Relevant.
Id. at 14.
As to Solers' rating, DISA now concedes that the NCES and
ASTG contracts were, in fact, related to cross domain
requirements. DISA Response to GAO Questions at 7-8. The
agency contends, however, that while the SRD states that
the two contracts "are not related [to] cross domain . . .
and hence the lower rating," the lower rating was not in
fact based on a finding that the two contracts were not
related to cross domain requirements. Id.; see AR, Tab L,
SRD, at 13. Instead, the agency contends that Solers'
relevancy rating was based on the agency's review of the
overall requirements of the PWS. DISA Response to GAO
Questions at 7-8. Specifically, DISA contends that Solers'
rating was based on the agency's view that the
requirements of the NCES and ASTG contracts were not of
the same magnitude and complexity as the requirements of
the PWS. Id.
We think that DISA's response to the protest is flatly
contradicted by the contemporaneous record. In this
regard, the record does not reflect that the agency
considered the overall relevance of the work; instead, the
evaluation specifically stated that the two contracts "are
not related [to] cross domain . . . and hence the lower
rating." AR, Tab L, SRD, at 13. To the extent that the
agency now contends that it actually meant that the NCES
and ASTF contracts did relate to cross domain requirements
and that the relevancy rating was based on a different
assessment, this post hoc argument does not demonstrate
that the agency's evaluation was reasonable. See Boeing
Sikorsky Aircraft Support, supra.
With regard to the evaluation of BAH's past performance
relevancy, the contemporaneous record does not discuss
which of BAH's past performance references were deemed
relevant, or what aspects of those references were
considered to involve the same effort as required by the
RFQ. In response to the protest, the agency states that it
relied on the personal knowledge of the evaluators in
concluding that BAH's past performance merited a
"substantial confidence rating." DISA Response to GAO
Questions at 8-9. As discussed above, however, there is no
discussion in the contemporaneous record concerning the
use of personal knowledge by the evaluators or what that
knowledge concerned.
Finally, the agency contends that the evaluators reviewed
the offerors' technical proposals to determine the
relevancy of their past performance, thus providing a
basis to make judgments concerning past performance
references for which there was no interview or PPIRS data.
DISA Response to GAO Questions at 7-9. Nothing in the
record supports this contention.[18] Instead, this
argument was first raised by the agency in response to
questions posed by our Office concerning the adequacy of
the record. Our Office accords greater weight to
contemporaneous source selection materials and documents
and little weight to arguments raised by counsel that are
not supported by the contemporaneous record. See Haworth,
Inc., B-297077, B‑297077.2, Nov. 23, 2005, 2005 CPD para.
215 at 8; Chemical Demilitarization Assocs., B‑277700,
Nov. 13, 1997, 98-1 CPD para. 171 at 10.
Past Performance Conclusion
In sum, we think that the record here is inadequate for
our Office to determine whether DISA's evaluation of the
offerors' past performance confidence and relevancy was
reasonable. In this regard, the contemporaneous record
does not document the information upon which the
evaluators appeared to have relied in making the
judgments, and the agency's response to the protest relies
on numerous assertions, which are either unsupported or
contradicted by the contemporaneous record. On this
record, we sustain the protest. (Solers,
Inc., B-404032.3; B-404032.4, April 6, 2011) (pdf)
TCDI
challenges the agency's evaluation of the firm's proposal
under a number of corporate experience/past performance
evaluation subfactors.
First, the protester complains that State unreasonably
assessed a weakness in TCDI's proposal under the past
performance subfactor, after concluding that a number of
TCDI's delivered products, specifically the WMEAT
database, CWC/MIMS, and the Treaty Information Portal,
required significant rework and/or reengineering. Protest
at 20. With respect to State's assertion that rework was
required on the WMEAT database, TCDI explains that it was
tasked with converting the database from an obsolete
version of Oracle to the Microsoft SQL Server database
system in 2007.[10] Afterwards, system tests revealed
errors in the Microsoft functions and not in TCDI's work.
TCDI explains that it reengineered the scripts to produce
the correct output and rewrote the computer code to enable
the system to work. Id. With respect to the agency's
reference to rework for CWC/MIMS, TCDI states that it is
unfamiliar with the acronym and speculates that it may
have been a project from 10‑12 years earlier. Id. n. 9.
With respect to the Treaty Information Portal, TCDI
disputes that the product required significant rework,
noting that the previous contracting officer's technical
representative was very satisfied with the work. Id. at
20.
In reviewing protests of alleged improper evaluations and
source selections, even in a task order competition as
here, we do not reevaluate proposals, but rather we
examine the record to determine whether the agency's
judgment was reasonable and in accord with the stated
evaluation criteria and applicable procurement laws and
regulations. ACCESS Sys. Inc., supra. In order for us to
review an agency's evaluation judgment, an agency must
have adequate documentation to support its judgment.
Northeast MEP Servs., Inc., B-285963.5 et al., Jan. 5,
2001, 2001 CPD para. 28 at 7. We do not limit our review
to contemporaneous evidence, but consider all the
information provided, including the parties' arguments,
explanations, and documentation prepared in response to
protest contentions. Systems Research and Applications
Corp.; Booz Allen Hamilton, Inc., B-299818 et al., Sept.
6, 2007, 2008 CPD para. 28 at 12. While we consider the
entire record, including the parties' later explanations
and arguments, we accord greater weight to contemporaneous
evaluation and source selection material than to arguments
and documentation prepared in response to protest
contentions. Id. Where an agency fails to provide
documentation of its evaluation, it bears the risk that
there may not be adequately supporting rationale in the
record for us to conclude the agency had a reasonable
basis for its evaluation and selection decision. Southwest
Marine, Inc.; American Sys. Eng'g Corp., B-265865.3,
B-265865.4, Jan. 23, 1996, 96-1 CPD para. 56 at 10.
State responds that its evaluation was based on "observed
performance and user feedback" but provides no explanation
or documentation of what was observed or what feedback it
received. See AR at 6. Although State contends that TCDI's
work on the WMEAT database was not timely completed, it
does not address TCDI's explanation as to why the firm was
not responsible for problems in the database. The agency
also does not address in any fashion TCDI's arguments with
respect to the assessed concerns with CWC/MIMS and the
Treaty Information Portal. In short, State has failed to
provide either contemporaneous documentation or subsequent
explanation supporting its assessment of this weakness in
TCDI's proposal. Accordingly, we find from the record
before us that State did not have a reasonable basis for
the assignment of this weakness in TCDI's proposal.
Next, TCDI challenges the agency's assessment of a
weakness under the corporate experience subfactor that
TCDI's proposal lacked "depth and breadth of experience
relevant to working within the Federal sector" as compared
to other offerors' proposals. Protest at 7-10. TCDI
disagrees with the agency's assessment and contends that
the RFP did not indicate that experience with other
government agencies was required. Id. at 9. TCDI states
that it drafted its proposal to focus on its experience
with State because this is what the solicitation indicated
would be evaluated. Comments at 24 n.13.
State responds that it considered this to be a minor
weakness and that the RFP "clearly indicated" that the
depth of work outside State would be considered because
the program requires interface with interagency partners.
AR at 5. However, the agency does not identify where in
the RFP offerors were notified that the depth, breadth,
and scope of an offeror's relevant work for other federal
agencies was a requirement. Moreover, we find that the
agency's evaluation record does not demonstrate that the
assignment of this weakness was reasonable. State does not
explain why TCDI's experience as the incumbent did not
provide sufficient experience working with other agencies,
such as Commerce and the Department of Defense.
TCDI also challenges State's evaluation of its proposal
under the personnel qualifications/management factor,
where it received a rating of green. Protest at 21.
Although the TEP noted that TCDI, as the incumbent, had
nearly all staff cleared and ready to immediately perform,
it also noted that TCDI had failed to propose any teaming
partner staff. Specifically, the TEP stated that, although
TCDI's proposal stated that the firm had sufficient access
to staff through its teaming partner to meet any perceived
shortfalls related to certification or expertise, TCDI did
not specifically offer any staffing from its teaming
partner. AR, Tab 4, TEP Evaluation Report, at 5. TCDI
argues that the RFP did not require offerors to identify
all proposed staff, and therefore the State Department
evaluated proposals on a different basis than that
announced in the solicitation. Protest at 24. Moreover,
TCDI contends that the agency's assignment of a weakness
in this area is irrational, given that the State
Department also noted a strength in TCDI's proposal under
the corporate quality certifications subfactor for the
various certifications of its teaming partner. Id. at 24
n.11.
State does not refute TCDI's arguments, nor does it
provide an explanation for the apparent inconsistency
between the weakness assigned under one evaluation factor
and the strength assigned in a subfactor. Rather, State
merely notes that TCDI's proposal was rated green under
this evaluation factor. AR at 7.
Here too, we have no basis to find State's evaluation to
be reasonable, given the agency's failure to provide any
documentation or support for its evaluation. In this
regard, we also agree with TCDI that the RFP did not
require offerors to identify staff other than key
personnel. (Technology
Concepts & Design, Inc., B-403949.2; B-403949.3, March
25, 2011 (pdf)
A1
Procurement complains that the VA improperly rejected its
proposal on the basis of the firm's low price without any
consideration of the firm's technical approach. In this
regard, A1 Procurement states it was able to offer its low
price because the firm's headquarters is only two blocks
from the cemetery. Protest at 5; Protester's Proposal at
6. The protester also complains that its proposed overall
price was only 6 percent below the awardee's price.
Comments at 8.
The VA responds that A1 Procurement's proposed price was
considered to be too low because the firm's proposed unit
price of $7,200 for the trimming CLIN was considered too
low in comparison to the government's unit price estimate
of $16,200 for this CLIN. The agency states that on this
basis it concluded that the protester's proposal was
technically unacceptable. Supp. AR at 7-8.
Before awarding a fixed-price contract, an agency is
required to determine that the price offered is fair and
reasonable. Federal Acquisition Regulation (FAR) sect.
15.402(a). An agency's concern in making a price
reasonableness determination focuses primarily on whether
the offered prices are higher than warranted, as opposed
to lower.[4] CSE Constr., B‑291268.2, Dec. 16, 2002, 2002
CPD para. 207 at 4. Although not required, an agency also
may provide for a price realism analysis in a solicitation
for the award of a fixed-price contract for the purpose of
assessing whether an offeror's low price reflected its
understanding of the contract requirements or the risk
inherent in an offeror's approach. Id.
Here, there is no explanation in the record as to the
purpose for the agency's evaluation of whether an
offeror's proposed fixed price was too low. The RFP does
not state that the VA intended to perform a price realism
analysis to assess the offerors' understanding or to
assess performance risk, nor does the VA assert that it
performed a price realism analysis.
Regardless of the agency's purpose for assessing whether
offerors' proposed fixed prices were too low, however, we
conclude that the VA's assessment that A1 Procurement's
price was too low is unsupported by the record and does
not appear to be consistent with the solicitation. As
noted above, the RFP provided that the agency would
consider, among other things, the offerors' price
breakdowns, pricing methodology, and technical approaches
and capabilities in assessing whether proposed prices were
too low or too high. RFP at 90. There is no documentation
in the record (nor does the agency claim) that such an
analysis was performed with respect to the protester's
proposal. Rather, the record contains only the VA's
conclusory judgment that the protester's fixed price was
too low. An agency's evaluation must be sufficiently
documented to allow review of the merits of a protest.
Northeast MEP Servs., Inc., B-285963.5 et al., Jan 5,
2001, 2001 CPD para. 28 at 7. Where an agency fails to
provide documentation of its evaluation, it bears the risk
that there may not be adequately supporting rationale in
the record for us to conclude the agency had a reasonable
basis for the selection decision. Id.
Furthermore, here the VA initially informed the protester
that its price was too low and represented a performance
risk because A1 Procurement's fixed price was
significantly below the GCE and the prices of the other
offerors. See AR, Tab 6, VA Debriefing Letter, Dec. 1,
2010. In response, the protester noted that its price was
only 6 percent below the awardee's price. Rather than
explain why it viewed the protester's overall fixed price
to be too low, the VA asserted that A1 Procurement's
proposal was rejected because the agency found the firm's
proposed price for one of the five CLINs to be too low.
See Supp. AR at 7-8. There is no contemporaneous
documentation evidencing the agency's concern that the
protester's price for one CLIN was too low in comparison
to the GCE for this CLIN. Nor does the VA state how the
protester's price compares to the awardee's or other
offerors' prices for this CLIN or overall. There is also
no explanation in the record as to why A1 Procurement's
low price for this one CLIN reflects a performance risk.
In short, the record is devoid of contemporaneous
documentation showing that the agency evaluated A1
Procurement's price in accordance with the RFP, and
contains no explanation in response to the protest to
support the agency's conclusion that A1 Procurement's
price is too low. Accordingly, we find unreasonable the
VA's rejection of A1 Procurement's proposal on the basis
of the agency's unsupported judgment that the firm's price
was too low. (A1
Procurement, JVG, B-404618, March 14, 2011) (pdf)
NEMS
contends, among other things, that the agency ignored
adverse past performance information concerning the
awardee's performance at a number of DeCA commissaries.
See Protest at 5; Comments at 9-12; Protester's Hearing
Comments at 6.
The record here shows that the technical evaluation team
received past performance information from DeCA
commissaries and from two other federal agencies. See AR,
Tab 9, Past Performance Information, at 92-185. Although
this information included a number of favorable comments
regarding Nayyarsons' past performance, e-mails from
several DeCA commissaries also indicated that Nayyarsons
had a number of performance problems. Specifically, the
evaluation team was provided with e-mails from various
commissaries which reported staff and product shortages
during transition periods; unsanitary conditions; employee
tardiness and cleanliness; and problems with sushi,
including use of expired products and pre-dating products.
See, e.g., id. at 147-48 (June 25 e-mail reporting
"serious problems" with sushi at Memphis commissary which
"nee[d] to be addressed immediately"); 145-46 (July 15
e-mails reporting an "ongoing problem" and that "there are
still significant issues" with sushi at Memphis
commissary); 154 (July 18 e-mail reporting problems "once
again" with sushi at Scott AFB).
As a general matter, the evaluation of an offeror's past
performance is within the discretion of the contracting
agency, and we will not substitute our judgment for
reasonably based past performance ratings. See MFM Lamey
Group, LLC, B‑402377, Mar. 25, 2010, 2010 CPD para. 81 at
10. While we have recognized that such judgments are often
subjective by nature, the exercise of these judgments in
the evaluation of proposals must be documented in
sufficient detail to show that they are not arbitrary.
That is to say, in order for us to review an agency's
evaluation of proposals, an agency must have adequate
documentation to support its judgment. Northeast MEP Servs.,
Inc., B‑285963.5 et al., Jan. 5, 2001, 2001 CPD para. 28
at 7. See American President Lines, Ltd., B-236834.3, July
20, 1990, 90-2 CPD para. 53 at 6.
Where a protester challenges the past performance
evaluation and source selection, we will review the
evaluation and award decision to determine if they were
reasonable and consistent with the solicitation's
evaluation scheme and procurement statutes and
regulations, and to ensure that the agency adequately
documented the basis for the selection. Wackenhut Servs.,
Inc., B-400240, B‑400240.2, Sept. 10, 2008, 2008 CPD para.
184 at 6; S4, Inc., B-299817, B‑299817.2, Aug. 23, 2007,
2007 CPD para. 164 at 9. When an agency fails to document
or retain evaluation materials, it bears the risk that
there may not be adequate supporting rationale in the
record for us to conclude that the agency had a reasonable
basis for its source selection decision. Southwest Marine,
Inc.; American Sys. Eng'g Corp., B‑265865.3, B-265865.4,
Jan. 23, 1996, 96-1 CPD para. 56 at 10.
The contemporaneous record here shows no consideration by
the technical evaluation team or the CO of the awardee's
recent performance problems identified in these e‑mails.
Rather, the evaluation team rated Nayyarsons' proposal
"exceptional" for each of the two past performance
subfactors, despite the team's receipt of numerous e-mail
reports of problems with Nayyarsons' performance. See AR,
Tab 9, Past Performance Information, at 89‑92. Moreover,
the record shows that several of the e‑mails identifying
problems with Nayyarsons' performance were originally
addressed to the CO here, and, during the hearing, the CO
testified that she reviewed all of the information that
was provided to the evaluation team. See, e.g., id. at
132, 145, 148, 150-54 (e-mail reports of awardee's adverse
past performance addressed to CO); Tr. at 74.
In our view, Nayyarsons' past performance rating is
inconsistent with the agency's stated rating scheme, which
provided that an "exceptional" rating would only be
assigned for past performance reflecting few minor
problems for which corrective actions taken by the
contractor were highly effective. As discussed above,
nothing in contemporaneous record shows that the agency
considered whether the firm in fact took effective steps
to correct the performance problems reported by DeCA
commissaries. See AR, Tab 9, Evaluators' Notes, at 75-91;
Tab 5, Decision Summary, at 1-23. For example, the record
shows that the Memphis commissary reported serious
problems with Nayyarsons' sushi products on June 25, but
that significant problems with sushi were ongoing as of
July 15. The record also shows that on July 18, the
commissary at Scott AFB reported renewed problems with
Nayyarsons' sushi products.
Based on this record, we conclude that the agency ignored
adverse past performance information and--in assigning a
past performance rating of exceptional to the awardee--deviated
from the agency's evaluation scheme. See G. Marine Diesel,
B‑232619.3, Aug. 3, 1989, 89-2 CPD para. 101 at 6‑7
(protest sustained where agency ignored awardee's
performance problems); Apptis, Inc., B‑299457, et al., May
23, 2007, 2008 CPD para. 49 at 11 (evaluation cannot be
determined to be reasonable where record lacks
documentation that agency considered deficiencies in
awardee's proposal); Midland Supply, Inc., B‑298720.3, May
14, 2007, 2007 CPD para. 104 at 5-6 (past performance
evaluation unreasonable and inconsistent with awardee's
record and agency's rating scheme); International Bus.
Sys., Inc., B‑275554, Mar. 3, 1997, 97‑1 CPD para. 114 at
5 (past performance information too close at hand for
agency to fail to consider the information). Accordingly,
we sustain the protest on this basis. (Northeast
Military Sales, Inc., B-404153,January 13, 2011)
(pdf)
Contrack argues that the agency's assessment of Zafer's
past performance as excellent was unreasonable because the
Corps failed to consider adverse past performance
information concerning Zafer's past performance.
Specifically, the protester complains that the agency did
not consider negative CPARs for three of Zafer's projects,
one of which was a contract with the Corps for the
renovation and repair of the National Army Military
Hospital in Kabul, Afghanistan. This contract was the
subject of the DoD IG report that Contrack provided to the
Corps with its initial protest.
Contrack also complains that the agency ignored the
news article (which was provided with the initial protest)
that reported that the DoD IG would be investigating other
deficiencies found by the Commission on Wartime
Contracting in Iraq and Afghanistan with respect to
Zafer's performance of another Corps contract for
construction services related to a headquarters building
in Kabul.[4] Protest at 11. Contrack contends that the
agency, in its review of the CPARs, unreasonably assigned
Zafer's proposal an excellent rating based upon the
company's performance of two smaller projects, while
ignoring negative reports for the three larger projects.
Comments at 7.
The Corps responds that Zafer's excellent past performance
rating was based upon the "totality of the past
performance information," which the Corps argues was
"laudatory." See CO's Statement at 2. In this regard, the
SSA states that one of the projects upon which the Corps
based its overall past performance rating was for an
ammunition supply point, which the Corps states was more
directly relevant than Zafer's other construction
projects. Finally, the Corps argues that it had no
obligation to consider the adverse information about
Zafer's prior performance contained in the DoD IG report,
any materials developed by the Wartime Contracting
Commission, or the news report concerning Zafer's past
performance. In the agency's view, the Corps acted
reasonably in ignoring these "external materials," even
though the agency acknowledges that these materials were
provided to it by Contrack as part of the earlier protest
filings. AR at 5-6.
As a general matter, the evaluation of an offeror's past
performance is within the discretion of the contracting
agency, and we will not substitute our judgment for
reasonably based past performance ratings. However, we
will question an agency's evaluation conclusions where
they are unreasonable or undocumented. Clean Harbors Envtl.
Servs., Inc., B-296176.2, Dec. 9, 2005, 2005 CPD para. 222
at 3; OSI Collection Servs., Inc., B-286597, B-286597.2,
Jan. 17, 2001, 2001 CPD para. 18 at 6. The critical
question is whether the evaluation was conducted fairly,
reasonably, and in accordance with the solicitation's
evaluation scheme, and whether it was based on relevant
information sufficient to make a reasonable determination
of the offeror's past performance, including relevant
information close at hand or known by the contracting
personnel awarding the contract. OSI Collection Servs.,
Inc., supra, at 6. As explained below, we find that the
agency's past performance evaluation did not meet this
standard.
The record shows that Zafer's excellent past performance
rating was primarily based upon two outstanding CCASS
performance ratings for projects in Afghanistan, which the
SSA found to be similar to the current requirement. See
Original CO's Statement, Oct. 2, 2009, at 2. The original
SSEB noted as strengths in Zafer's proposal that Zafer had
provided many outstanding letters of commendation, that
there were minimal field changes and requests for
equitable adjustments, that Zafer had outstanding ratings
on projects in Afghanistan, that Zafer had a good safety
record, and that many of Zafer's projects were completed
ahead of schedule. See AR (B‑401871), Tab 8, Original SSEB
Report, at 18-19.
As noted by Contrack during its initial protest, Zafer's
initial proposal identified nine projects for its
experience and, in accordance with the RFP requirements,
provided past performance forms for all nine projects. See
Protest at 11. The Corps obtained CPARs for five of
Zafer's identified projects (all of which were performed
in Afghanistan), which the Corps provided with its report
on Contrack's initial protest. Three of these reports,
however, provided overall ratings of satisfactory or
marginal, and indicated that Zafer had a number of
performance problems on these projects. For example with
respect to the CPAR for Zafer's National Army Military
Hospital project in Kabul (overall value of $18.9
million), the report indicated that Zafer's performance
rating was marginal overall; the report also contained
such remarks as "[d]eficient work was not corrected in a
timely [manner]," "multiple notices of deficiency were
issued with no resolution," and "[p]erformance was
consistently poor without correction." This was the
project that was the subject of the DoD IG report that
Contrack provided to the Corps with its initial protest.
See Protester's Comments, exh. 5, CPAR for Contract No.
W917PM-05-C-0005. Similarly, the CPAR for another of
Zafer's reported projects, the design and build of a
hospital facility at Baghram Airfield, indicated that
Zafer's performance received an overall satisfactory
rating, but noted that some of its work (including the
quality of its work) was assessed as marginal.
Although the Corps generally argues that it based its
assessment of Zafer's past performance as excellent upon
the "totality" of Zafer's record, there is no evidence in
the contemporaneous evaluation record of the agency's
consideration of these three CPARs, which were identified
by the protester. Moreover, the Corps has not shown in
response to the protest how it could reasonably conclude
that Zafer's overall past performance was excellent in
light of the three CPARs indicating only satisfactory or
marginal performance on a number of construction projects
in Afghanistan. In this regard, we recognize that the
Corps may not have been aware, at the time of its initial
evaluation and selection decision, of the DoD IG report
and other information that may indicate negative past
performance by Zafer, but the Corps was provided with this
information during the earlier protest. Nonetheless, there
is no indication in the record that the Corps made any
effort to investigate the merits of these reports, at
least one of which appears to be documented in one of the
CPARs available to the agency. In fact, it appears from
the record that the newly‑appointed SSEB only reviewed
Zafer's revised proposal under the past performance factor
to see if the firm addressed the single weakness
identified during discussions. See AR, Tab 10, SSEB
Report, Jan. 26, 2010, at 3; see also Tab 10, SSEB
Consensus Evaluation Worksheet for Zafer, Past Performance
Factor ("no increase in rating – already Excellent").
In the absence of evidence in the record that the Corps
considered relevant information available to it concerning
Zafer's past performance, we have no basis to find
reasonable the agency's assessment of an excellent rating
for Zafer under the past performance factor. We sustain
Contrack's protest on this basis. See GTS Duratek, Inc.,
B-280511.2, B-280511.3, Oct. 19, 1998, 98-2 CPD para. 130
at 14 (protest sustained where agency ignored known past
performance information). (Contrack
International, Inc., B-401871.5, B-401871.6,
B-401871.7, May 24, 2010) (pdf)
The ATO and DEA protests raise numerous issues regarding
the Navy's public-private competition and the selection of
PMI. The protesters argue, among other things, that the
Navy's evaluation of PMI's technical and price proposals
was unreasonable, that the agency's discussions with the
ATO were not meaningful, and that the Navy's corrective
action improperly limited the aspects of proposals that
offerors could revise. The protesters also argue that the
contract award to PMI is improper because the Navy failed
to complete the public-private competition study here
within the required 30-month statutory timeframe regarding
the expenditure of appropriated funds for such studies.
As detailed below, we find the agency's evaluation of
PMI's staffing plan--a protest ground raised only by the
ATO--was improper. Although we do not specifically address
the protesters' remaining challenges to the Navy's
evaluation of proposals, we have fully considered all of
them and find that they are without merit.
An agency that fails to adequately document its evaluation
of proposals bears the risk that its determinations will
be considered unsupported, and absent such support, our
Office may be unable to determine whether the agency had a
reasonable basis for its determinations. Rosemary
Livingston--Agency Tender Official, B-401102.2, July 6,
2009, 2009 CPD para. 135 at 10, recon. denied, Department
of the Navy--Request for Modification of Remedy,
B-401102.3, Aug. 6, 2009, 2009 CPD para. 162 at 4; Rhonda
Podojil--Agency Tender Official, B-311310, May 9, 2008,
2008 CPD para. 94 at 4; Matrix Int'l Logistics, Inc.,
B‑272388.2, Dec. 9, 1996, 97-2 CPD para. 89 at 5.
In this case, as explained above, the Navy's original
staffing estimate for the mobile refueling fuel sampling
and testing requirement (Spec Item 3.3.1) was predicated
on the historical number of samples completed (5,824) and
an estimated 40 minutes per occurrence, which (together
with overtime) resulted in a total of 3,946 labor hours.
When the Navy later changed its estimate to 30 minutes per
occurrence, the Navy then concluded that 2,956 total labor
hours were required to adequately perform mobile refueler
fuel sampling and testing requirement.
The TEB determined that PMI's original proposal with
regard to the staffing of Spec Item 3.3.1 was deficient,
insofar as the 9.9 minutes per fuel sampling and testing
occurrence was too low. PMI's response to discussions
revised the fuel sampling and testing time upward to 15.88
minutes, but the offeror also reduced the number of fuel
sampling and testing occurrences to 2,720. However, PMI's
proposal did not explain how it had determined that the
PWS required only 2,720 fuel samples annually,[11] or how
it had determined that 15.88 minutes per occurrence was
adequate to perform the requirements; it also appears that
the offeror "backed into" its fuel sampling time by first
allocating the total number of labor hours.
The TEB subsequently concluded that PMI's revised proposal
had remedied all deficiencies and that its proposed
staffing for Spec Item 3.3.1 (and ELIN 005) was adequate.
However, the record does not indicate why it was
reasonable for the TEB to conclude that PMI's proposed
time of 15.88 minutes per fuel sample test was adequate,
when it differed substantially from the Navy's time
estimate. The record also does not indicate why it was
reasonable for the TEB to conclude that PMI's assumption
that only 2,720 fuel test samples per year were required
was acceptable, when it differed substantially from the
agency's IGE and the historical number of 5,824 annual
fuel test samples. As a result, the record does not
support the TEB's conclusion that PMI's proposed 960 labor
hours were sufficient to perform Spec Item 3.3.1, given
that the Navy believed an offeror's staffing had to be
based on at least 2,956 total hours.
Further, the record shows that when the ATO attempted to
revise its staffing for Spec Item 3.3.1 by using a
15-minute sampling time, the TEB expressly disagreed and
found that it made the ATO's proposal technically
unacceptable.[12] In this regard, while the TEB found that
the ATO's proposal had failed to adequately support how
the MEO would accomplish the fuel tests in 15 minutes per
sample, the record shows that PMI's proposal--which the
TEB found acceptable--likewise failed to provide any
information to support how it would accomplish the same
tests in 15.88 minutes per sample, referring only to the
increase from its original, lower time per sample.
Similarly, with regard to the number of fuel samples
required, the record shows that while the TEB found that
PMI's assumption and the corresponding staffing were
acceptable, the TEB used the higher IGE number (5,824
samples) and corresponding staffing when assessing the
adequacy of the ATO's proposed staffing for the
requirement. In sum, while the TEB evaluated the ATO's
proposal based on a required minimum of 2,956 labor hours
for Spec Item 3.3.1 (and concluded that the 1,529 labor
hours proposed by the ATO were unacceptable), the TEB
concluded without explanation that the 960 labor hours
proposed by PMI for the same task were sufficient. There
is no explanation in the record for the inconsistencies in
the agency's evaluation of PMI's and the ATO's proposals
in this area.
The agency argues that it reasonably determined the
adequacy of PMI's staffing for all of ELIN 0005, and that
it was at the ELIN-level (and not the underlying Spec
Item-level) that staffing adequacy was measured. The
record does not support this argument. As a preliminary
matter, the record reflects that the IGE had staffing
estimates for each Spec Item, and that the ELIN staffing
estimates were based on the Spec Item staffing estimates.
The record also reflects that the Navy assessed the
adequacy of offerors' staffing plans at both the Spec Item
and ELIN levels. Moreover, when determining that the ATO's
staffing plan for Spec Item 3.3.1 was insufficient, the
TEB's evaluation did not extend beyond the one specific
task and consider the offeror's staffing for the entire
ELIN. The agency also found PMI's original staffing for
Spec Item 3.3.1 to be a deficiency, and there is nothing
in the record to suggest that it was remedied by PMI's
staffing of other aspects of ELIN 0005.
The Navy also argues that the evaluation record was
adequate to support its determination that PMI's staffing
for Spec Item 3.3.1 was acceptable; the agency essentially
argues that "not much need be said" when determining a
proposal meets (as opposed to failing to meet) the
requirements. We disagree. An agency's evaluation of
proposals must be adequately documented in order to
establish the reasonableness of its determinations. See
Urban-Meridian Joint Venture, B-287168, B-287168.2, May 7,
2001, 2001 CPD para. 91 at 2. This requirement applies
equally to evaluation determinations of proposal
acceptability and determinations of proposal
unacceptability, weakness, or deficiency. Here, the TEB
concluded without explanation or other support that PMI's
revised proposal had adequately staffed Spec Item 3.3.1
notwithstanding the fact that the underlying number of
fuel sample occurrences, time per occurrence, and labor
hours, differed materially from the Navy's estimates. To
the extent some other aspect of PMI's proposal made its
staffing for the mobile refueler fuel sampling and testing
requirement acceptable (e.g., cross-utilization of labor
from other Spec Items), such analysis is not part of the
TEB's conclusory determination of PMI's staffing adequacy.
Given the inadequate documentation in the record before us
to support a key finding that PMI's staffing plan was
acceptable, we sustain the protest on this basis. (Bruce
Bancroft--Agency Tender Official; Sam
Rodriquez--Designated Employee Agent, B-400404.7;
B-400404.8; B-400404.9; B-400404.10; B-400404.11, November
17, 2009) (pdf)
Our Office will review the documentation supporting a
source selection decision to determine whether that
decision was adequately supported and rationally related
to the evaluation factors. Universal Shipping Co.,
B-223905.2, Apr. 20, 1987, 87-1 CPD para. 424, at 10.
Implicit in the foregoing is that the evaluation must be
documented in sufficient detail to show that it was not
arbitrary, Adelaide Blomfield Mgmt. Co., B-253128.2, Sept.
27, 1993, 93-2 CPD para. 197 at 4, and a selection
decision may not be made on point scores alone where the
agency selection official has inadequate documentation on
which to base a reasoned decision. J.A. Jones Mgmt. Servs.,
Inc., B‑276864, July 24, 1997, 97-2 CPD para. 47 at 4.
While we are mindful that when an agency places an order
under a BPA, limited documentation of the source selection
is permissible, the agency must at least provide a
sufficient record to show that the source selection was
reasonable. FAR sect. 13.303-5(e); see also FAR sections
13.106‑3(b)(3)(ii) (in a simplified acquisition record
must include additional statements to explain the basis
for award when factors other than price are used).
In short, the contemporaneous record here is inadequate
for our Office to determine whether the Forest Service had
a reasonable basis to select Aquatic at its higher price.
Additionally, the record shows that the Forest service
used an evaluation factor (equipment) that was not
disclosed in the solicitation, and which had the effect of
reducing the significance of the cost/price factor.[7]
Since the record provides no contemporaneous tradeoff
comparing Aquatic to C&B, other than on the basis of their
point scores, we sustain the protest. See. Shumaker
Trucking & Excavating Contractors, Inc., B-290732, Sept.
25, 2002, 2002 CPD para. 169 at 8 (protest sustained where
Forest Service relied solely on point scores and failed to
document any comparison of protester's lower-priced and
lower-rated proposal to awardee's higher-priced,
higher-rated proposal, in source selection decision).
With respect to the Forest Service's more detailed
explanation of its evaluation, offered in response to
C&B's protests, our Office generally gives little weight
to such reevaluations and judgments prepared in the heat
of the adversarial process. Boeing Sikorsky Aircraft
Support, B‑277263.2, B‑277263.3, Sept. 29, 1997, 97‑2 CPD
para. 91 at 15. In our view, the explanation offered by
the Forest Service in response to the protests cannot
overcome the failure of the contemporaneous record to
provide any explanation of the evaluation of C&B (other
than point scores) on which the agency could make a
reasoned selection decision.
(C&B Construction, Inc.,
B-401988.2, January 6, 2010) (pdf)
Finally,
GAI protests that the agency’s evaluation and source
selection decision was inadequately documented,
complaining that “[n]o evaluation worksheets, individual
evaluator notes or scales were produced.” GAI Comments and
Supplemental Protest, Mar. 5, 2009, at 3. GAI’s protest in
this regard is without merit.
Although an agency must document its evaluation judgments
in sufficient detail to show that they are not arbitrary,
the necessary amount and level of detail will vary from
procurement to procurement. U.S. Defense Sys., Inc.,
B-245563, Jan. 17, 1992, 92-1 CPD para. 89 at 3;
Champion-Alliance, Inc., B-249504, Dec. 1, 1992, 92-2 CPD
para. 386 at 6-7. For example, there is no requirement
that the evaluation record must include narrative
explanations for every rating assigned. Apex Marine Ship
Mgmt. Co., LLC; American V-Ships Marine, Ltd.,
B-278276.25, B-278276.28, Sept. 25, 2000, 2000 CPD para.
164 at 8-9. Similarly, there is no requirement that an
agency retain individual evaluator’s notes or worksheets,
provided the agency’s final evaluation documentation
reasonably explains the basis for the agency’s judgments.
Global Eng’g and Constr. LLC, B-290288.3, B‑290288.4, Apr.
3, 2003, 2003 CPD para. 180 at 3 n.3.
Here, the contemporaneous record included a detailed
technical evaluation report that included the agency’s
narrative explanation regarding the basis for each
evaluation rating of either “Unacceptable” or
“Outstanding.” AR, Tab 32. Further, the narrative
explanations supporting the ratings consistently include
specific references to the particular portions of the
offerors’ proposals that formed the basis for the agency’s
assessments. Id. Finally, the agency’s source selection
documentation contained a detailed comparative discussion
of the two offerors’ proposals, identifying particular
distinguishing features of each proposal. AR, Tab 34. On
this record there is no merit in GAI’s assertion that the
agency’s evaluation and source selection decision were
inadequately documented. (Government
Acquisitions, Inc., B-401048; B-401048.2; B-401048.3, May
4, 2009) (pdf)
Here, the
record is inadequate to establish that the contracting
officer’s finding of technical equality is reasonable and
proper. As discussed above, the evaluation record consists
of the evaluators’ adjectival ratings for each of the
subcriteria, their narrative comments under several
subcriteria, and the contracting officer’s scoring of the
proposals based on the adjectival ratings. The evaluators
did not provide the contracting officer with a
comprehensive assessment or listing of the proposals’
strengths and weaknesses, and the record includes no
evidence that the contracting officer ever considered the
actual merits of the proposals in calculating the scores.
Likewise, there is no indication that the contracting
officer considered the actual merits of the proposals in
ultimately determining that, notwithstanding ROG’s
proposal’s approximately 12 percent higher score, it was
equal in technical merit to RPCI’s proposal. The record
includes no explanation of the contracting officer’s
rationale for her conclusion that the approximate 12
percent scoring difference did not translate into actual
technical superiority for RPG’s proposal. Rather, the
record includes only the conclusory statement that “After
performing the evaluation, it was determined that the
offers were equal both technically and in past performance
… .” AR exh. 15, Price Negotiation Memorandum, Aug. 23,
2007. This brief statement is the sole contemporaneous
explanation for the contracting officer’s determination
that the proposals were technically equal, notwithstanding
ROG’s proposal’s higher percentage score.
In reviewing an agency’s evaluation, we may also consider
documentation prepared after the source selection decision
was made, although we will accord greater weight to
contemporaneous materials rather than judgments made in
response to protest contentions. Beacon Auto Parts,
B-287483, June 13, 2001, 2001 CPD para. 116 at 6. Here,
the agency submitted no contracting officer’s statement in
response to the protester’s supplemental protest, in which
ROG’s specific evaluation challenges are raised, and the
post-protest record, like the contemporaneous record,
contains no other support for the contracting officer’s
conclusion that the proposals were technically equal. In a
memorandum dated October 11, 2007, prepared after the
filing of RPCI’s prior protest, the contracting officer
concluded that “The combined scoring for Technical and
Past Performance was found to be equal.” AR exh. 14,
Contracting Officer Memorandum, Oct. 11, 2007. Similarly,
in the narrative submitted in response to ROG’s initial
protest, the contracting officer merely recites that she
“determined that the offerors were essentially technically
equal based on the technical and past performance factors
set forth in the RFP.” AR exh. 1, Contracting Officer’s
Narrative, July 7, 2008, at 1. These conclusory statements
are inadequate to establish the reasonableness of the
contracting officer’s determination that the proposals
were technically equal.
While the agency’s report in response to ROG’s
supplemental protest does not include a statement by the
contracting officer, it does respond to each of the
protester’s specific challenges to the evaluation ratings.
However, these responses were provided by the agency’s
legal counsel, with no indication that the responses
reflect the contracting officer’s own rationale for her
evaluation conclusions. In this regard, the agency’s
counsel provides explanations for the various challenged
ratings but, instead of attributing the asserted
rationales to the contracting officer, asserts that the
explanations would lead “a reasonable person” to conclude
that the scoring was reasonable. Supp. AR at 5-7. These
responses do not constitute an adequate evaluation record,
see York Bldg. Servs., Inc., B‑296948.2 et al., Nov. 3,
2005, 2005 CPD para. 202 at 7 (GAO accords little or no
weight to “new rationales, based on a hypothetically
correct evaluation, for which there is no support in the
contemporaneous record.”), and the supplemental report
does not otherwise indicate the considerations that
factored into the contracting officer’s determination that
the proposals were technically equal.
We conclude that the contracting officer’s determination
that ROG’s and RPCI’s proposals were technically equal
lacked adequate supporting explanation or documentation
and, therefore, was unreasonable. See Magellan Health
Servs., B‑298912, Jan. 5, 2007, 2007 CPD para. 81 (protest
challenging adequacy of agency’s source selection decision
sustained where evaluation record was insufficient to
establish reasonableness of the selection official’s
determination that offers were technically equal,
notwithstanding protester’s proposal’s higher technical
rating); Midland Supply, Inc., B‑298720, B‑298720.2, Nov.
29, 2006, 2007 CPD para. 2 (award decision not reasonable
where there is no documentation or explanation and agency
makes its award decision based strictly on a mechanical
comparison of the offerors’ total point scores)Accordingly,
we sustain the protest on this ground.
Radiation Oncology Group of WNY,
PC, B-310354.2; B-310354.3, September 18, 2008) (pdf)
The agency
tender official asserts that, after finding the agency
tender "acceptable" after several rounds of discussions,
the agency improperly conducted yet another round of
discussions, as a result of which the agency tender no
longer was the lowest-priced tender/offer received. In
response, the agency argues that the TEB in fact did not
find the agency tender acceptable until after the final
round of discussions, despite having described the agency
tender as "acceptable" in the TEB report prepared after
the prior round of discussions. The key question in
resolving the protest thus is whether the agency's
evaluation record adequately shows that the agency did--or
did not--find the agency tender acceptable before the
final round of discussions. As discussed below, we think
the record is inconclusive in this regard.
Although the FAR does not specify what is required to be
documented in the contract file in support of an agency’s
evaluation of proposals, see FAR sect. 15.305(a), the
fundamental principle of government accountability
dictates that an agency maintain a record adequate to
allow for the meaningful review of the merits of a
protest. This principle applies equally in the context of
a public-private competition under the Circular. See
Rhonda Podojil--Agency Tender Official, B-311310, May 9,
2008, 2008 CPD para. 94 at 4. An agency that fails to
adequately document its evaluation of proposals bears the
risk that its determinations will be considered
unsupported, and absent such support, our Office may be
unable to determine whether the agency had a reasonable
basis for its determinations. Matrix Int’l Logistics,
Inc., B‑272388.2, Dec. 9, 1996, 97-2 CPD para. 89 at 5.
That is not to say that our Office, in determining the
reasonableness of an agency’s evaluation and award
decision, limits its review to the contemporaneous
evaluation and source selection documentation. Rather, we
will consider, in addition to the contemporaneous
documentation, all information provided to our Office for
consideration during the protest, including the parties'
arguments and explanations, so long as the information is
credible and consistent with the contemporaneous record.
Id.; NWT, Inc.; PharmChem Labs., Inc., B-280988,
B-280988.2, Dec. 17, 1998, 98-2 CPD para. 158 at 16.
In considering the entire record, we accord greater weight
to contemporaneous evaluation and source selection
material than to the parties' later explanations,
arguments, and testimony. Boeing Sikorsky Aircraft
Support, B-277263.2, B-277263.3, Sept. 29, 1997, 97-2 CPD
para. 91 at 15. Where the record before us is inconsistent
or incomplete to such an extent that we cannot find the
agency's evaluation of proposals to be reasonable, we will
sustain the protest. Carahsoft Tech. Corp.; Allied Tech.
Group, B-311241, B-311241.2, May 16, 2008, 2008 CPD para.
119 at 8-9 (sustaining protest where the record contained
inadequate documentation to show the reasonableness of the
agency's evaluation, and the agency's arguments appeared
inconsistent with the contemporaneous record); Honeywell
Tech. Solutions, Inc.; Wyle Labs., Inc., B‑292354,
B-292388, Sept. 2, 2003, 2005 CPD para. 107 at 7
(sustaining protest where "we simply cannot determine from
[the] record which aspect of the agency’s evaluation was
reasonable and which was unreasonable").
In this case, as explained above, the agency held a total
of five rounds of discussions with the agency tender
official; at the conclusion of the fourth round, the
agency tender was lower-priced than the eventual awardee's
offer. Thus, if (as the agency tender official argues) the
agency tender in fact was found acceptable after the
fourth round of discussions, holding the fifth and final
round of discussions was improper. Further, those
discussions resulted in prejudice to the agency tender
because the agency tender's price increased as a result,
to a price greater than the eventual awardee's, thus
displacing the agency tender as the lowest-priced
technically acceptable offer/tender. In response, the
agency argues that the record clearly shows that, despite
having described the agency tender as "acceptable" before
initiating the fifth round of discussions, the TEB in fact
found the agency tender "unacceptable," and thus properly
conducted another round of discussions.
As discussed below, we think that the evaluation record is
inconsistent and inconclusive with respect to the TEB's
findings regarding the agency tender after the fourth
round of discussions. As a result, we conclude that the
record here is inadequate to support a conclusion that the
decision to hold the final round of discussions was
proper. Accordingly, we sustain the protest on this basis.
The dispute regarding the TEB's findings derives
principally from the conflicting language in the July TEB
report, specifically, the repeated description of the
agency tender as "acceptable" alongside this sentence:
"This weakness will need to be corrected before
implementing the MEO can be considered." The agency does
not assert that these statements are reconcilable; rather,
the agency argues that the TEB made an error in describing
the agency tender as "acceptable" and that the quoted
sentence from the TEB report is contemporaneous evidence
that the TEB in fact considered the fourth revised agency
tender "unacceptable." In support of its position, the
agency, pointing to the RFP definitions of "acceptable"
and "marginal" proposals, asserts that the agency tender
could not be regarded as "acceptable" because it contained
a weakness that had to be corrected before implementation
of the MEO.
Even accepting the agency's interpretation of the RFP
definitions of "acceptable" and "marginal," the agency's
argument still does not resolve the conflict between the
TEB report's description of the agency tender as
"acceptable" and the finding in the same report that the
agency tender had a weakness needing correction before the
MEO could be considered. Those two statements cannot both
be accurate, and we see no basis to conclude that the
"mistake" was in the characterization of the agency tender
as "acceptable," rather than in the finding that the
agency tender contained a deficiency, as evidenced by the
statement that the agency tender contained a weakness
requiring correction. On the contrary, the record strongly
suggests that the description of the agency tender as
"acceptable" reflects a deliberate choice by the TEB,
given that, in the prior three rounds of discussions and
evaluations, the agency characterized the agency tender as
"poor" before making the significant change--to repeatedly
describing the agency tender as "acceptable," sometimes in
bold capital letters--in the July report.
The agency also argues that the discussions letter to the
agency tender official identifying deficiencies in the
tender, which stated that, "[a]lthough your tender remains
in the competitive range, it continues to contain
deficiencies," is contemporaneous evidence that the TEB
had found the then-most recent tender to be unacceptable.
We are not persuaded that this letter reasonably can be
regarded as reflecting a determination by the agency that
the agency tender was "unacceptable." As noted above, the
language of that letter is essentially identical to the
language used in the prior discussions letter, suggesting
that the agency may have copied the "boilerplate" portion
of the prior letter rather made a deliberate language
choice. Similarly, the letter uses the plural
term--"deficiencies"--to describe the agency tender, while
the TEB Report at most identified a single deficiency in
the agency tender, further suggesting that the letter does
not warrant the dispositive weight urged by the agency.
Finally, the agency argues that, notwithstanding that the
agency tender was mistakenly labeled "acceptable" in the
July 29, 2008 TEB report, it is otherwise clear from the
record that the fourth revised agency tender failed to
meet a material term of the RFP and therefore was not
acceptable. As noted above, that TEB report stated that
the agency tender had the “appropriate amount of [DELETED]
FTEs to handle [DELETED]; however, the TEB is still
concerned that the [DELETED] FTEs will not be able to
handle [DELETED],” AR, Exh. 31, TEB Report of July 29,
2008 at 25, and it was not until its response to the
agency tender official’s comments on the agency report
that the agency first alleged that the revised agency
tender failed to meet a government requirement. The TEB
found that the agency tender had met the more
time-intensive requirement of the RFP with respect to
[DELETED], that is, it had staffed [DELETED]. Now the TEB
was left to determine whether the agency tender had
allotted sufficient hours [DELETED]. The TEB report
recounted in some detail the rationale that the agency
tender advanced in support of its assertion that it had
allotted sufficient hours, without concluding that the
agency tender was inadequate in any way. A full reading of
the TEB report suggests that the TEB was identifying a
concern about the hours allocated [DELETED], but that,
given the approach to the requirement proposed in the
agency tender, the agency tender was no longer
unacceptable or deficient. Absent a clear statement that
the fourth revised tender failed to meet a material
requirement of the RFP--statements that the agency
previously had made repeatedly throughout the
competition--on this record we do not think it is
reasonable to infer such a finding, particularly where
doing so would contradict the agency's own explicit
statement that the tender was "acceptable."
On the record before us, with inconsistent statements by
the agency in its evaluation of the fourth revised tender
and inadequate documentation of a finding that the tender
was unacceptable, we sustain the protest. (Rosemary
Livingston--Agency Tender Official, B-401102.2, July
6, 2009) (pdf) See (Department
of the Navy--Request for Modification of Remedy,
B-401102.3, August 6, 2009) (pdf)
MINACT
asserts that the evaluators' contemporaneous reasons for
not increasing its score are invalid and fail to provide a
rational basis for the evaluation. TRESP Assocs, Inc.;
Advanced Data Concepts, B‑258322.5, B‑258322.6, Mar. 9,
1995, 96‑1 CPD para. 8 at 4. It further asserts that the
evaluators' hearing explanations are both irrational and
represent post hoc rationalizations, to be accorded
little, if any weight. Post Hearing Comments at 27; see
Boeing Sikorsky Aircraft Support, B‑277263.2, B-277263.3,
Sept. 29, 1997, 97-2 CPD para. 91 at 15.
An agency is required to document its judgments in
sufficient detail to show that they are not arbitrary.
TRESP Assocs, Inc.; Advanced Data Concepts, supra.
However, in reviewing an agency's evaluation, we do not
limit our review to contemporaneous evidence, but consider
all of the information provided, including the parties'
arguments, explanations, and any hearing testimony.
Remington Arms Co., Inc., B-297374, B‑297374.2, Jan. 12,
2006, 2006 CPD para. 32 at 10. While we generally give
little weight to reevaluations and judgments prepared in
the heat of the adversarial process, Boeing Sikorsky
Aircraft Support, supra, post‑protest explanations that
provide a detailed rationale for contemporaneous
conclusions simply fill in previously unrecorded details,
and will generally be considered in our review of the
reasonableness of an evaluation, so long as those
explanations are credible and consistent with the
contemporaneous record. NWT, Inc.; PharmChem Labs., Inc.,
B‑280988, B-280988.2, Dec. 17, 1998, 98-2 CPD para. 158 at
16.
The evaluators' explanations are sufficient to establish a
reasonable basis for their evaluations. In this regard,
MINACT's argument is based upon a faulty assumption--that
its proposal was downgraded based on the assessed
weaknesses and that, when the weaknesses were corrected,
its proposal score should have been increased by all
remaining points--in essence, that it should have received
a perfect score. Post Hearing Comments at 28; exh. 27, at
15. There is no basis for this assumption. First, nothing
in the RFP indicated that more than a satisfactory score
would be assigned proposals meeting the RFP requirements.
Absent such an RFP provision, we think it is indisputably
reasonable for a proposal to be rated with an adjective of
fully satisfactory--rather than very good or
exceptional--where the correction of weaknesses results in
the proposal's meeting--rather than exceeding--the RFP
requirements. This was the approach followed by the
evaluators. Tr. at 50, 151, 209. Since, as explained by
the evaluators, MINACT's improvements merely brought its
proposal up to the expected and satisfactory level, as set
forth in the RFP, and were found to provide no value
beyond that level, the evaluators could reasonably
conclude that the elimination of the identified weaknesses
did not warrant a rating above satisfactory. Likewise, we
have no basis to question the evaluators' conclusion that
the addition of a single strength in 1 of 11 areas
reviewed did not warrant an increase in MINACT's score.
See Tr. at 155. Certainly, MINACT has not established that
the correction of weaknesses and the addition of a single
strength warranted a perfect score. (MINACT,
Inc., B-400951, March 27, 2009) (pdf)
The protester alleges that the agency improperly focused its
evaluation on the final year of the 3 years of past performance
that should have been considered, and that the agency has not
produced the necessary contemporaneous documentation to support
the score [DELETED] given the protester. As a general matter,
the evaluation of an offeror’s past performance is a matter
within the discretion of the contracting agency, and we will not
substitute our judgment for reasonably based past performance
ratings. In determining whether a particular evaluation
conclusion is reasonable, we examine the record to determine
whether the judgment was reasonable and in accord with the
evaluation criteria listed in the solicitation. Abt Assocs.,
Inc., B-237060.2, Feb. 26, 1990, 90-1 CPD para. 223 at 4.
Implicit in that examination is that the evaluation must be
documented in sufficient detail to show that it was not
arbitrary. Federal Acquisition Regulation (FAR) sections
15.305(a), 15.308; Quality Elevator Co. Inc., B-276750, July 23,
1997, 97‑2 CPD para. 28 at 3.
According to the agency, the TET members relied primarily on an
oral discussion of their individual experiences with HTS in
arriving at the past performance rating of [DELETED].
Supplemental AR at xviii. There is no contemporaneous
documentation of that discussion. In an evaluation that takes
into account the agency’s own knowledge of offerors’
performance, the fundamental requirement that evaluation
judgments be documented in sufficient detail to show that they
are reasonable and not arbitrary must still be met. Omega World
Travel, Inc., B‑271262.2, July 25, 1996, 96-2 CPD para. 44 at 4.
The agency argues that evidence of HTS's past performance, first
introduced into the record as part of the agency report on the
protest, supports the reasonableness of the evaluation, citing
Omega World Travel, Inc.. In that case, however, the record
showed that the agency evaluators relied in their deliberations
on specific evaluation material subsequently produced in the
protest record. Here, there is no contemporaneous account of the
discussion of HTS's past performance, let alone one that
references the evaluators' reliance on any of the written
documentation the agency supplied in response to the protest.
The TET chair asserts that the team based its evaluation of the
protester's past performance on "the written evaluations that it
had," Declaration of TET Chair at para.15, but fails to identify
any specific document or to state whether the evaluators
considered any of the contracts listed in the protester's
proposal. In contrast, the TET chair goes into considerable
detail, in response to the protest, recounting the personal
experiences of the TET members that contributed to the rating.
Id. at paras. 6‑14. The documents produced in the agency report
may or may not be part of the written evaluations that the TET
asserts it relied on, but the TET chair's recollection of the
evaluation process indicates that greater weight was given to
the undocumented team discussion of past performance.
As discussed above, evaluations must be documented in sufficient
detail to show that they were not arbitrary. FAR sections
15.305(a), 15.308. Here, without any contemporaneous
documentation of that oral evaluation, and with no record,
contemporaneous or otherwise, of what contract performance
information was considered and how much relevance the
information was accorded, we have no basis on which to conclude
that the past performance evaluation was reasonable.
(Helicopter Transport Services
LLC, B-400295; B-400295.2, September 29, 2008) (pdf)
In reviewing
protests of alleged improper evaluations and source selection
decisions, it is not our role to reevaluate submissions; rather,
we will examine the record to determine whether the agency’s
judgment was reasonable and in accord with the stated evaluation
criteria and applicable procurement laws and regulations.
However, for our Office to perform a meaningful review, the
record must contain adequate documentation showing the bases for
the evaluation conclusions and source selection decision.
Panacea Consulting, Inc., B-299307.4, B- 299308.4, July 27,
2007, 2007 CPD para. 141 at 3-4. Where an agency fails to
document or retain evaluation materials, it bears the risk that
there may not be adequate supporting rationale in the record for
us to conclude that the agency had a reasonable basis for the
source selection decision. Systems Research & Applications
Corp.; Booz Allen Hamilton, Inc., B-299818 et al.
Sept. 6, 2007, 2008 CPD para. 28 at 12.
In our view, the record here is inadequately documented to show
the reasonableness of GSA’s decision to exclude compensation
management from the protesters’ awards. Additionally, GSA’s
arguments in this protest appear inconsistent with the
contemporaneous record of the TEP conclusions. While the TEP’s
conclusions are fairly general, they do not assign the
protesters a “high risk,” a “red” rating, or any other rating
indicating technical unacceptability for any aspect of
compensation management. In addition, there is no record that
the contracting officer arrived at a reasoned independent
judgment regarding rating, risk, or acceptability of the
protesters’ proposals under the RFP criteria at any time after
the December 18 TEP report, which assigned a rating of “yellow”
and “moderate risk” to these proposals. The TEP report itself is
otherwise silent on the basis for implicitly excluding
compensation management from the awards.
In attempting to address the lack of any discussion of how the
TEP evaluated the protesters’ approach to compensation
management, GSA argued that “[s]ince there was no demonstration
of the protesters’ ability to perform using GSA supplied data,
there was nothing for the TEP to evaluate or discuss with
respect to the award of Compensation Management.” Supp. AR at
2-3.
We think GSA’s argument is not supported by the record. In the
RFP, and again in the additional instructions provided before
the demonstration, GSA sought significant information about how
offerors would provide compensation management services. Very
simply, while it is undisputed that the protesters did not
demonstrate the Ceridian product using GSA’s sample test data,
both protesters provided information on their compensation
management approach in their written proposals, and demonstrated
the functions of their private payroll processing subcontractor
using commercial data. In light of this, we fail to see a
reasonable basis for GSA’s position that there was nothing for
the TEP to evaluate, and, in fact, the contemporaneous summary
ratings given in the final TEP report suggest that GSA did not
find the protesters’ approach unacceptable. Moreover, we think
the argument that these proposals provided nothing to discuss or
evaluate is both inaccurate and unfair, and is belied by the
record.
In summary, we conclude that GSA’s decision to exclude
compensation management services from the protesters’ awards is
not reasonably supported by either the contemporaneous record or
the agency’s explanations during this protest. Accordingly, we
need not reach the specific issues raised by the protesters
concerning the evaluation of their proposals and the meaning of
specific provisions in the RFP. (Carahsoft
Technology Corporation; Allied Technology Group, B-311241;
B-311241.2, May 16, 2008) (pdf)
KC-30 Overrun and Breakaway Capability
Boeing also complains that the Air Force did not reasonably
assess the capability of Northrop Grumman’s proposed aircraft to
refuel all current Air Force fixed-wing tanker-compatible
aircraft using current Air Force procedures, as required by a
KPP No. 1 threshold under the aerial refueling area of the key
system requirements subfactor. See RFP, SRD sect. 3.2.10.1.1.9.
Specifically, Boeing notes that current Air Force refueling
procedures require that the tanker aircraft be capable of
“overrun” and “breakaway” procedures when necessary, which would
require the tanker aircraft to fly faster than the receiver
aircraft or quickly accelerate during refueling. Boeing’s Second
Supplemental Protest at 29. Boeing contends that the Air Force
unreasonably determined that Northrop Grumman’s proposed
aircraft would meet these requirements.
With regard to the overrun issue, the record shows that Northrop
Grumman was twice informed by the Air Force during discussions
that the firm’s initially identified maximum operational
airspeed of [Deleted] Knots Indicated Air Speed (KIAS) would not
be sufficient under current Air Force overrun procedures to
achieve required overrun speeds of [Deleted] KIAS for various
fighter aircraft, including the [Deleted], or [Deleted] KIAS for
the [Deleted]. See AR, Tab 184, EN NPG-MC1-003, at 2; EN
NPG-MC1-003a, at 1-2. Ultimately, Northrop Grumman informed the
Air Force that a [Deleted] limited the aircraft’s operational
speed, but that Northrop Grumman proposed to include a [Deleted]
to achieve the necessary overrun speed. See id., Northrop
Grumman Response to EN NPG‑MC1‑003a, at 2-7. The Air Force
accepted Northrop Grumman’s proposed solution as satisfying this
KPP threshold. HT at 628.
Boeing complains that Northrop Grumman’s proposed solution of
[Deleted] to achieve overrun speed requires [Deleted], which is
not consistent with the Air Force’s current procedures as is
required by the KPP. See Boeing’s Second Supplemental Protest at
29-32; Boeing’s Comments at 64. Boeing also argues that the
agency did not note that Northrop Grumman qualified its promise
to increase its maximum operational airspeed in its EN response.
Specifically, Boeing points out that Northrop Grumman stated
that, [Deleted], the KC‑30 had a maximum airspeed of [Deleted]
KIAS, and not the [Deleted] KIAS evaluated by the Air Force. See
AR, Tab 184, Northrop Grumman Response to EN NPG-MC1-003a, at 9.
At the hearing that our Office conducted in this protest, the
Air Force produced its SSET mission capability factor team chief
to testify regarding the agency’s evaluation of the capability
of Northrop Grumman’s aircraft to satisfy this KPP threshold.
This witness, in response to direct examination, stated that the
SSET found that [Deleted] would allow the KC-30 to achieve the
necessary airspeed to perform the required overrun and breakaway
procedures. Specifically, he testified that the SSET was
convinced that, by [Deleted], the KC-30 could achieve an
operational airspeed of [Deleted] KIAS, because Northrop Grumman
had informed the agency in its EN response that the commercial
A330 aircraft had a maximum “dive velocity” of 365 KIAS and had
been flight tested to a dive velocity of [Deleted] KIAS, and
that analysis had been done showing that the A330 could even
achieve a dive velocity of [Deleted] KIAS. HT at 626-27. The
mission capability factor team chief testified that the SSET
evaluated Northrop Grumman’s response to indicate that the
[Deleted], see HT at 637-38, and that in any event Air Force
current procedures did not require the use of the [Deleted]
during aerial refueling operations. HT at 638-39.
From this record, we cannot conclude that the Air Force
reasonably evaluated the capability of Northrop Grumman’s
proposed aircraft to satisfy the KPP threshold requirement to
refuel all current Air Force fixed-wing tanker-compatible
aircraft using current Air Force procedures. The contemporaneous
record, as explained by the hearing testimony, does not
establish that the Air Force understood Northrop Grumman’s
response in discussions concerning its ability to satisfy the
solicitation requirements, nor does it demonstrate that the
agency had a reasonable basis upon which to accept Northrop
Grumman’s promises of compliance.
First, we agree with Boeing that the SSET erred in concluding
that the [Deleted] in tanker refueling operations was not a
current Air Force procedure. See HT at 638, 735; Air Force’s
Post-Hearing Comments at 19. As noted above, the contemporaneous
evaluation record shows that the agency interpreted the
solicitation requirement to comply with “current [Air Force]
procedures” to mean compliance with the procedures set forth in
the agency’s flight manuals for the KC‑135 and KC-10 tanker
aircraft, and expressly informed Northrop Grumman during
discussions that the flight manuals for the KC-135 and KC-10
established the current Air Force procedures for refueling
operations. See AR, Tab 184, EN NPG‑MC1-003a, at 1, wherein the
agency stated “[a]erial refueling procedures were contained in
T.O. 1-1C-1-3 and 1-1C-1-33 for the KC-135 and KC-10
respectively when the RFP was released.”[62] These manuals show
that current Air Force procedures provide that tanker pilots
[Deleted] in refueling operations. For example, the KC-135
manual under Section IV, Air Refueling Procedures, warns tanker
pilots that they “must be prepared to assume aircraft control
[Deleted],” and under Section V, Emergency Air Refueling
Procedures, instructs tanker pilots that in a breakaway
situation, if a climb is required, they must “[Deleted].” See
AR, Tab 289, Flight Manual KC-135 (Tanker) Flight Crew Air
Refueling Procedures, Supp. III, T.O. 1-1C-1-3, Jan. 1, 1987, as
revised Sept. 1, 2004, at [Deleted]. Similarly, the KC‑10 flight
manual provides under Section III, Air Refueling Procedures,
that the “[Deleted].” Id., Flight Manual, KC-10A Aircraft,
Flight Crew Tanker Air Refueling Procedures, USAF Series, T.O.
1‑1C-1-33, Sept. 1, 2002, as revised Jan. 31, 2005, at
[Deleted]. In this regard, Boeing provided the statement of a
retired Air Force pilot, who had extensive experience as both a
KC-10 and KC-135 tanker pilot and had operated each aircraft as
both a tanker and a receiver in refueling missions; this
individual stated:
Refueling is more demanding and difficult for both tanker and
receiver aircraft if the tanker [Deleted]. For the tanker
pilot, [Deleted]. For the receiver pilot, [Deleted]. Due to
these realities, existing refueling guidelines dictate that
[Deleted] should be used for refueling under normal
circumstances. [Citations omitted.] Beginning aerial refueling
[Deleted] should it become necessary, violates this policy. As
previously noted, [Deleted].
Boeing’s Comments, attach. 14, Declaration of Retired Air Force
Pilot, at 3-4. Although the Air Force and Northrop Grumman
generally disagree with Boeing’s consultant that the Air Force’s
current procedures provide for the [Deleted], neither the agency
or intervenor have directed our attention to anything in the
KC‑135 or KC‑10 flight manuals or to any other source that would
establish that Boeing’s view, which appears to be reasonable on
its face, is in error.
We also find unsupported the agency’s conclusion that Northrop
Grumman’s proposed solution of [Deleted] did not also involve
[Deleted]. In its EN response, Northrop Grumman informed the Air
Force that 330 KIAS was the normal design maximum operating
velocity of the commercial A330 aircraft, and that “selection of
a [maximum operating velocity] drives overall design
characteristics of the aircraft, specifically aerodynamic and
structural design limits, handling quality definition, and
thrust.” See AR, Tab 184, Northrop Grumman Response to EN NPG‑MC1-003a,
at 2. Northrop Grumman explained that its [Deleted] limited the
aircraft to its maximum operating velocity, but that the firm
could [Deleted] to exceed the maximum operating velocity. The
awardee then stated “three cases . . . to illustrate the
performance of the KC-30 with and without [Deleted].” Id. at 3.
The three cases that Northrop Grumman identified and separately
described were (1) KC-30 [Deleted]; (2) KC-30 [Deleted]; and (3)
KC-30 [Deleted], which indicated that the KC-30 could only meet
the overrun requirement under the third case where both the
[Deleted]. Id. at 3-6.
The SSET read, as described by the testimony of its mission
capability factor team chief, Northrop Grumman’s EN response to
describe a “fourth case” (although not identified as such) under
the “third case” heading, but located at the end of that
section, where, the agency contends, the KC‑30’s [Deleted] but
the [Deleted]. See HT at 664. However, we are unable to accept
such a reading of Northrop Grumman’s EN response. It ignores the
logical structure of Northrop Grumman’s response to the agency,
which only identified and described three cases. Moreover,
nowhere in its response to the agency’s EN does Northrop Grumman
suggest a “fourth case” where the [Deleted]; rather, the only
reference to both the [Deleted] in the third case expressly
states that the [Deleted] (“Case 3: KC-30 [Deleted]”).[63] See
AR, Tab 184, Northrop Grumman Response to EN NPG‑MC1-003a, at 6.
In any event, given the uncertainty surrounding the agency’s
interpretation of Northrop Grumman’s solution to a matter the
agency believed could render the firm’s proposal unacceptable,
see HT at 625, 649, this is something the agency should have
continued to clarify and resolve during discussions with the
firm.
Even apart from the agency’s apparent misreading of Northrop
Grumman’s EN response and disregard of the current Air Force
procedure to [Deleted], the record does not establish that the
agency had a reasonable basis for concluding that Northrop
Grumman’s proposed solution would allow its aircraft to obtain
the requisite overrun airspeeds to satisfy this KPP threshold.
The witness that the Air Force produced to support its arguments
on this point testified that the SSET had concluded that the
KC-30 had the “inherent capability” of reaching airspeeds
greater than [Deleted] KIAS (the aircraft’s certified maximum
operational airspeed) based upon the far greater airspeed
([Deleted] KIAS) identified by the firm for its certified dive
velocity.[65] See HT at 624-28; Air Force’s Post-Hearing
Comments at 17‑18. In this regard, the SSET apparently believed
that simply [Deleted] would enable the aircraft to achieve its
indicated dive velocity airspeed as its operational airspeed.
Although the SSET mission capability factor team chief
repeatedly testified that the dive speed indicated that the
aircraft would have the structural ability to fly at the dive
speed limitation, see, e.g., HT at 674, he also admitted under
cross examination that he did not know what the relationship was
between maximum operating airspeed and design dive speed:
Q: What’s your understanding of what the general margin is
between maximum operational velocity and dive velocity?
A: I’m not aware.
Q: Was there somebody on your team that was advising you about
what the general margin is or difference is between maximum
operational velocity and dive velocity?
A: There could have been. We had advisors for handling
qualities.
Q: I know you had advisors. I’m asking you, were there any
advisors who actually helped you with understanding the
difference between dive velocity and maximum operational
velocity?
A: They did not help me, no.
Q: Did they help the team?
A: Not that I’m aware of.
HT at 669-70. The SSET mission capability factor team chief’s
(and presumably the SSET’s) lack of knowledge concerning the
relationship between maximum operating airspeed and design dive
airspeed[66] is particularly troubling given the definition of
maximum operating limit speed in FAA’s regulations:
The maximum operating limit speed . . . is a speed that may
not be deliberately exceeded in any regime of flight (climb,
cruise, or descent), unless a higher speed is authorized for
flight test or pilot training operations. [The maximum
operating limit speed] must be established so that it is not
greater than the design cruising speed . . . and so that it is
sufficiently below [dive speed and velocity] to make it highly
improbable that the latter speeds will be inadvertently
exceeded in operations.
14 C.F.R. sect. 25.1505.
In sum, despite having identified, as an issue for the hearing,
the capability of Northrop Grumman’s proposed aircraft to
satisfy the airspeed requirements of this KPP threshold, we have
been presented with no testimony or documented analysis that
explains why simply [Deleted] on the KC-30 would ensure that the
proposed aircraft would achieve required overrun airspeeds that
are in excess of its FAA certified maximum airspeed.[67]
Furthermore, neither the Air Force nor Northrop Grumman has
directed us to any documentation establishing that the agency
analyzed what would be entailed in designing the KC-30 to exceed
the certified maximum operational airspeed limit.[68] Given
Northrop Grumman’s recognition in its EN response that selection
of the maximum operational airspeed limit “drives overall design
characteristics of the aircraft, specifically the aerodynamic
and structural design limits, handling quality definition, and
thrust,” see AR, Tab 184, Northrop Grumman Response to EN
NPG-MC1-003a, at 2, it would seem apparent that some design and
FAA re-certification efforts could be necessary.
Boeing also complains that the Air Force did not reasonably
evaluate the capability of Northrop Grumman’s aircraft to
initiate emergency breakaway procedures when refueling the
[Deleted]. Current Air Force procedures, as reflected by the
KC-135 flight manual, specifies that the tanker will refuel the
[Deleted] at an airspeed of [Deleted] KIAS, see AR, Tab 289,
Flight Manual KC‑135 (Tanker) Flight Crew Air Refueling
Procedures, Supp. III, T.O. 1-1C-1-3, Jan. 1, 1987, as revised
Sept. 1, 2004, at [Deleted], and Northrop Grumman’s EN response
indicates that the KC‑30’s airspeed is limited to [Deleted] KIAS
with the aircraft’s [Deleted]. See AR, Tab 184, Northrop Grumman
Response to EN NPG-MC1-003a, at 9. Boeing contends, citing the
statement of its former tanker/receiver pilot consultant, that
there is insufficient margin between airspeed at which [Deleted]
are refueled and the KC-30’s operational airspeed limit during
refueling (a [Deleted]-KIAS margin) to allow for emergency
breakaway maneuvers. See Boeing’s Comments, attach. 14,
Declaration of Retired Air Force Pilot, at 3-4.
As was true with respect to whether the KC-30 can satisfy the
current Air Force procedures with respect to overrun airspeed,
there is no documentation in the record setting forth an
analysis of whether Northrop Grumman’s proposed aircraft has
sufficient operational airspeed when refueling the [Deleted] to
initiate an emergency breakaway procedure. The agency’s counsel
provided a proffer at the hearing that the SSET’s analysis of
whether the KC-30 was capable of performing a breakaway maneuver
with the [Deleted] was contained in the SSET’s Final Evaluation
Summary Report for Northrop Grumman. See HT at 784; see AR, Tab
215, Evaluation Summary Report for Northrop Grumman, at 3.
Neither the page referenced by agency counsel or any other part
of that document contains any analysis of whether Northrop
Grumman’s proposed aircraft can perform a breakaway procedure
while refueling the [Deleted]; rather, the page referenced by
agency counsel merely states that “[t]he Offeror has
substantiated the ability to deliver a KC-X aircraft that meets
(minimum requirement) all KPP thresholds associated with aerial
refueling,” and provides no reasons or analysis supporting this
conclusion. AR, Tab 215, Evaluation Summary Report for Northrop
Grumman, at 3.
Although the SSET mission capability factor team chief was
examined extensively about the SSET’s consideration of the
KC-30’s ability to perform breakaway procedures, he recalled
little about the SSET’s discussions in this regard. His
testimony does indicate, however, that the SSET accepted that
the KC-30’s maximum operational airspeed when refueling
([Deleted]) was [Deleted] KIAS, and that the SSET apparently
believed that, to initiate the emergency breakaway procedure,
with Northrop Grumman’s proposed aircraft, the tanker would have
to start accelerating and [Deleted] simultaneously. See HT at
706. During cross examination, the SSET mission capability
factor team chief admitted that he did not know how long it
would take [Deleted] Northrop Grumman’s proposed [Deleted] or
what the procedure was for [Deleted], nor was he aware of
whether this was ever analyzed by the agency in its evaluation.
HT at 685-87, 707.
In sum, we conclude that the record does not demonstrate that
the agency reasonably determined that Northrop Grumman’s
proposed aircraft would be able to refuel all current Air Force
fixed-wing tanker-compatible receiver aircraft in accordance
with current Air Force procedures as was required by this KPP
No. 1 threshold.
(The
Boeing Company, B-311344; B-311344.3; B-311344.4;
B-311344.6; B-311344.7; B-311344.8; B-311344.10; B-311344.11,
June 18, 2008) (pdf)
ATS asserts that the best value determination was flawed because
the agency has provided no documents showing that proposals were
evaluated by individual evaluators and because the agency has
provided no evidence that it conducted a comparative analysis of
the proposals to determine which was the best value. This
argument is without merit. First, the lack of documents prepared
by individual evaluators does not render an agency’s evaluation
unreasonable per se; rather, we consider the record adequate if
the consensus documents and source selection decision
sufficiently document the agency’s rationale for the evaluation.
Joint Mgmt. and Tech. Servs., B-294229, B-294229.2, Sept. 22,
2004, 2004 CPD para. 208 at 3-4; Global Eng’g and Constr., LLC,
B-290288.3, B-290288.4, Apr. 3, 2003, 2003 CPD para. 180 at 3
n.3. Here, the evaluation documentation is sufficient because it
includes the evaluators’ consensus report, which details the
strengths and weaknesses of the proposals that formed the basis
for both the agency’s evaluation ratings for each offeror and
the selection decision itself. With respect to the absence of a
detailed comparative evaluation of the proposals, since the
proposals selected for award were both higher technically rated
and lower priced than ATS’s proposal, such a comparative
evaluation--i.e., a price-technical tradeoff--was not required.
MD Helicopters, Inc., Agusta Westland, Inc., B-298503 et al.,
Oct. 23, 2006, 2006 CPD para. 164 at 49 n.49.
(Alliance Technical Services, Inc.,
B-311329; B-311329.2, May 30, 2008) (pdf)
As noted above, the solicitation expressly provided that the
agency would evaluate the realism of the offerors’ proposed
prices, and that such analysis would include consideration of an
offeror’s [deleted]. RFP at 86.
Further, the solicitation provided that the agency would also
assess proposal risk, including an assessment of whether an
offeror’s proposed approach has potential for “disruption of
schedule, increased cost, degradation of performance, and the
need for increased Government oversight, as well as the
likelihood of unsuccessful contract performance.” Id. at
82.
It appears beyond reasonable dispute that, even in the context
of a fixed-price contract, an offeror’s proposed approach that
[deleted] could create risks of [deleted].
In order for our Office to meaningfully review an agency’s
evaluation, the agency must have adequate documentation to
support its various judgments. Where an agency fails to create
or retain documentation regarding its evaluation assessments, it
bears the risk that our Office will be unable to determine
whether the agency’s judgments were reasonable. Southwest
Marine, Inc.; American Sys. Eng’g Corp., B-265865.3,
B-265865.4, Jan. 23, 1996, 96-1 CPD ¶ 56 at 10.
Here, the agency’s evaluation record includes no meaningful
documentation addressing the unexplained changes in Boeing’s
assumptions between submission of its initial proposal and its
subsequent proposal revisions. Specifically, the record contains
no documentation regarding any agency consideration of the basis
for Boeing’s changed [deletion], how Boeing’s revised [deleted]
correspond to the reality of the [deleted], how the revised
[deleted] correspond to the agency’s own [deleted] projections,
or whether Boeing’s revised [deleted] are likely to create
[deleted]. Accordingly, on the record here, we are unable to
determine whether the agency reasonably concluded that Boeing’s
proposed price is realistic, or whether the agency’s assessment
of “low risk” for Boeing’s proposal, under each of the mission
capability subfactors, is reasonable in light of Boeing’s
revised [deleted]. Since we are unable to determine whether the
agency reasonably performed a price realism analysis, or
properly considered the potential risk flowing from Boeing’s
revised [deleted], we sustain the protest on this basis. (Pemco
Aeroplex, Inc., B-310372, December 27, 2007) (pdf)
In order for us to review an agency’s evaluation of proposals,
an agency must have adequate documentation to support its
judgment. Northeast MEP Servs., Inc., B-285963.5 et al., Jan. 5,
2001, 2001 CPD para. 28 at 7. While an agency is not required to
retain every document or worksheet generated during its
evaluation of proposals, the agency’s evaluation must be
sufficiently documented to allow review of the merits of a
protest. KMS Fusion, Inc., B-242529, May 8, 1991, 91-1 CPD para.
447 at 10. Where an agency fails to document or retain
evaluation materials, it bears the risk that there may not be
adequate supporting rationale in the record for us to conclude
that the agency had a reasonable basis for its source selection
decision. Southwest Marine, Inc.; American Sys. Eng’g Corp.,
B-265865.3, B-265865.4, Jan. 23, 1996, 96-1 CPD para. 56 at 10.
Here, the agency’s evaluation record contains the individual
evaluator comment reports (over 600 pages), as well as various
handwritten notes of SSEB members regarding Apptis’s POC
demonstration. AR, Tab 9C, SSEB Evaluation Comments of Apptis;
Tab 12, SSEB Notes from Apptis POC demonstration. However, none
of these contemporaneous documents contains any reference to
either of the nonsubstantive deficiencies that the evaluators
found in Apptis’s demonstration. Likewise, there are no other
documents in the evaluation record that support the SSEB’s
conclusions that Apptis’s personnel had repeated difficulty in
getting proposed solutions to work correctly the first time they
were demonstrated, or that each of the various components of
Apptis’s proposed solution required a different technical expert
to implement. Quite simply, while the agency described Apptis’s
POC as a “problem plagued demonstration,” AR, Tab 14, SSAC
Briefing to the SSA, at 23, it kept no records of which Apptis
solutions required multiple attempts, how many attempts were
required, or any reasons for the multiple attempts. In sum, we
cannot tell if the evaluation of this aspect of Apptis’s
proposal was reasonable because the agency record lacks adequate
documentation to support its findings regarding Apptis’s POC
demonstration. In its report to our Office, the agency
acknowledges that, “[t]here is no documentation from the POC
that identifies how many times it took [Apptis] to successfully
demonstrate any aspect of [its] proposed solutions.” AR, Apr. 5,
2007, at 9. DISA argues, however, that because the POC
demonstration was not separately evaluated, there was no reason
for the SSEB to document its concerns regarding Apptis’s
nonsubstantive deficiencies. The agency’s argument is based on a
fundamental misunderstanding of the requirement that agencies
provide an adequate basis for their evaluation findings. While
the POC demonstration was not itself a separate evaluation
factor, it was a significant part of the agency’s overall
evaluation of an offeror’s proposal and, as such, the agency was
required to maintain records adequate to permit meaningful
review. Tiger Enters., Inc., B-293951, July 26, 2004, 2004 CPD
para. 141 at 2. (Apptis,
Inc., B-299457; B-299457.2; B-299457.3, May 23, 2007) (pdf)
Here, notwithstanding the agency’s stated understanding as to
what its corrective action would entail, the fact is that the
record provided to our Office in response to the protest does
not contain sufficient information to support the evaluations or
source selection decisions. Under solicitation 446, for example,
the record includes a description of the agency’s method for
scoring the submissions under the understanding of the SOW,
relevant experience, capabilities, and approach evaluation
factor. The description states that the agency prepared a matrix
that included standards for the award of point scores ranging
from 0 points (for an unsatisfactory submission) to 10 points
(for an excellent submission), and states further that the
evaluators populated this matrix with the raw point scores
assigned to the submissions under five subfactors (such as the
vendor’s understanding of the SOW, and the vendor’s approach to
staff retention). Protest, exh. 1. This explanation goes on to
state, mathematically, how the agency calculated the ultimate
scores for the price and non-price factors. Id. However, the
record does not include any explanation of the bases for
assigning the raw point scores in the first place; that is,
there is no indication of, for example, any evaluated strengths
or weaknesses in the proposals that support the assigned scores.
The record does include one evaluator’s worksheets prepared in
connection with solicitation 446. Agency Report (AR) exh. 12.
The worksheets include some limited comments, but the comments
do not relate the evaluator’s conclusions about the submissions
to the scores assigned. Further, the worksheets reflect the
views of only one of the three evaluators who participated in
the evaluation and, thus, in no way can serve as a substitute
for a narrative explanation of the consensus scores assigned to
the submissions. Protest exh. 2. Moreover, the agency states
that, since this evaluator’s worksheets were prepared during the
original evaluation, they are irrelevant to the current
evaluation. Agency Supplemental Submission, June 15, 2007, at 2.
The evaluation materials for solicitation 448 are similarly
inadequate. As is the case with the other solicitation, the
record contains a document describing the scoring matrix, which
identifies the standards for assigning point scores of 0 to 10
as well as various subfactors used in the scoring. Protest exh.
3. Also, as with the other solicitation, the record fails to
explain what features of the submissions led the agency to score
them as it did. As with solicitation 446, the agency submitted
some individual scoresheets for solicitation 448. AR exhs. 9,
11. However, all of these scoresheets were prepared by a single
evaluator and contain only brief, often cryptic, notations
regarding the basis for the scores assigned. In all of the
scoresheets for solicitation 448, the portion of the scoresheet
relating to the understanding of the SOW, relevant experience,
capabilities and approach evaluation factor is entirely blank,
containing neither point scores nor narrative explanation. Id.
Further, there are several scoring anomalies that are nowhere
explained. For example, the agency assigned one of SID’s
employees a score of 13 points under a subfactor with a maximum
of 10 points available, and also assigned both a score of 5 and
a score of 3 (which were added together) under another subfactor,
again without any explanation. AR exh. 9, SID Scoresheet, at 4;
exh. 11, SID Scoresheet, at 4. The record also contains no
source selection decision document for either solicitation
prepared in connection with the agency’s corrective action.
Point scores cannot be used as a substitute for adequate
documentation showing the bases for the evaluation conclusions
reached and source selection decisions made. OSI Collection
Servs., Inc., supra, at 8. We conclude that the record contains
insufficient documentation to explain the basis for the
evaluations and the source selection decisions; accordingly, we
sustain this aspect of Panacea’s protest. (Panacea
Consulting, Inc., B-299307.4; B-299308.4, July 27, 2007) (pdf)
Here, it is clear that the critical criterion for assessing how
many additional linguists will be needed to successfully meet
the task order 1 fill rate requirements for annual productive
hours is the amount of time an offeror’s linguists are expected
to be in a non-productive status. Accordingly, it appears the
only logical basis for the agency’s conclusion that GLS’s
proposal of [deleted] linguists is more likely to meet the fill
rate requirements than L-3’s proposal of [deleted] linguists
would be a rationally supported conclusion that GLS’s [deleted]
linguists are likely to spend [deleted] less time in a
non-productive status than are L‑3’s [deleted] linguists. The
record, however, is devoid of any meaningful analysis addressing
that issue. Indeed, at the GAO hearing, the MET chair testified
that the agency gave no consideration to whether the additional
personnel proposed by GLS will be sufficient to compensate for
its linguists’ scheduled holidays and sick leave. Tr. at 526.
Consistent with this testimony the record reflects no agency
consideration of whether GLS’s linguists will spend more or less
time in a non-productive status, either scheduled or
unscheduled, than L-3’s linguists--nor has the agency
represented that such analysis was performed. Absent the
agency’s consideration of this issue, reasonably documented and
rationally supported by credible data, we are unable to conclude
that the agency reasonably evaluated L-3’s proposal of [deleted]
linguists as being more likely to create a shortfall against the
required fill rate than GLS’s proposal of [deleted] linguists. (L-3
Communications Titan Corporation, B-299317; B-299317.2;
B-299317.3, March 29, 2007) (pdf)
In order
for our Office to perform a meaningful review of an agency’s
award determination, the agency is required to have adequate
documentation to support its evaluation of proposals and its
award decision. Century Envtl. Hygiene, Inc., B-279378, June 5,
1998, 98-1 CPD para. 164 at 4; Biospherics, Inc., B-278508.4 et
al., Oct. 6, 1998, 98-2 CPD para. 96 at 4; Arco Mgmt. of Wash.,
D.C., Inc., B‑248653, Sept. 11, 1992, 92-2 CPD para. 173 at 3.
An award decision is not reasonable where there is no
documentation or explanation to support the price/technical
tradeoff and where the agency makes its award decision based
strictly on a mechanical comparison of the offerors’ total point
scores. Universal Bldg. Maint., Inc., B-282456, July 15, 1999,
99-2 CPD para. 32 at 4; see also FAR sections 12.602(c), 15.308.
The evaluation record here consists of one‑paragraph summaries
of the proposals, and charts showing for each evaluation factor
the agency’s assignment of raw scores, the calculation of
weighted scores, and the total point score for each proposal.
The record lacks any documentation reflecting a meaningful
comparative analysis of proposals or any explanation of why
Danaher’s lower technically rated, lower-priced proposal was
selected for award over Midland’s higher technically rated,
higher‑priced proposal. The record shows that the agency relied
on a mechanical comparison of the total point scores assigned to
the Midland and Danaher proposals without any qualitative
assessment of the technical differences between these proposals
to determine whether Midland’s technical superiority would
justify the payment of a price premium. (Midland
Supply, Inc., B-298720; B-298720.2, November 29, 2006) (pdf)
We have reviewed the record here and conclude that the
evaluation is not adequately supported. We note at the outset
that the evaluation record here is brief, comprised only of the
initial evaluation scoring sheets prepared by the individual
evaluators (for example, the record does not include the
consensus source selection evaluation report contemplated by the
source selection plan, AR, exh. 103, at 8); a brief summary of
advantages and disadvantages observed during the operational
demonstrations; and a brief source selection document. (In this
regard, the advantages and disadvantages observed during the
operational demonstration are identical to the advantages and
disadvantages included in the source selection document.) We
find that the evaluation judgments are, in many instances,
either factually incorrect, internally contradictory, or so
cryptic that we are unable to discern either the basis for the
evaluators’ concerns or how their concerns related to the
solicitation’s evaluation criteria. (Intercon
Associates, Inc., B-298282; B-298282.2, August 10, 2006) (pdf)
A consensus rating need not be the same as the rating
initially assigned by the individual evaluators; rather, the
final evaluation rating may be arrived at after discussions
among the evaluators. I.S. Grupe, Inc., B-278839, Mar. 20, 1998,
98-1 CPD ¶ 86 at 5-6. Where, as here, the agency has destroyed
individual evaluation materials, its actions are unobjectionable
provided that the consensus evaluation materials relied on by
the agency support the agency’s judgments regarding the relative
merits of the proposals. Global Eng’g and Constr., LLC,
B-290288.3, B-290288.4, Apr. 3, 2003, 2003 CPD ¶ 180 at 3 n.3.
We find no merit to this aspect of JMTS’s protest. As noted, the
record includes the agency’s consensus evaluation materials
which, contrary to JMTS’s assertion, provide a significant level
of detail about the evaluators’ findings regarding the strengths
and weaknesses identified in the proposals. While JMTS devotes a
significant portion of its protest to asserting that the
evaluators’ conclusions are erroneous primarily as they relate
to JMTS’s proposal, the protester’s disagreement with the
evaluation conclusions does not demonstrate that they are
lacking in detail. (Joint Management &
Technology Services, B-294229; B-294229.2, September 22,
2004) (pdf)
The record shows that the Air Force did not evaluate
Adelphia’s “record of integrity and business ethics” as part of
its past performance evaluation, as was specifically required by
the RFP. The Air Force does not assert that it performed such an
evaluation, but merely argues that its “very good” rating of
Adephia’s past performance was justified, given the information
provided and reviewed. The record shows that the Air Force
assessed Adelphia’s past performance as “very good” based only
upon the survey responses it received from the three contract
sources identified by Adelphia in its proposal and the DCMA
pre-award survey. See Agency Report, Tab 9, Past Performance
Assessment, at 1; Contracting Officer’s Statement at 3. The past
performance surveys, however, did not seek or receive any
information concerning Adelphia’s record of integrity or ethics.
[6] Furthermore, as asserted by the protester, Adelphia
performed these three referenced contracts while the indicted
members of the Rigas family had significant ownership interest
and control in the awardee and its parent companies. Given the
specific RFP language, the charges brought by the SEC and the
indictment of the Rigas family members should have been (but
were not) evaluated as relevant information in the agency’s
assessment of Adelphia’s past performance. In short, because the
agency’s past performance evaluation was not in accord with the
stated RFP criteria, we find the agency’s evaluation of
Adelphia’s past performance to be inconsistent with the RFP
evaluation scheme and unreasonable. See Beneco Enters., Inc., B-
283512.3, July 10, 2000, 2000 CPD ¶ 176 at 7.
We also find no basis in the record to find reasonable the Air
Force’s evaluation of Southwestern Bell’s past performance. As
indicated above, the agency’s entire explanation of its
evaluation rating of the protester’s past performance was that
Southwestern Bell had a dozen contracts with the Air Force’s
procuring office, that these contracts were “well documented,”
and that the protester’s evaluated price was higher than that of
the awardee. Agency Report, Tab 9, Past Performance Assessment,
at 2. Despite the agency’s statement that Southwestern Bell’s
performance of contracts with its office were “well documented,”
no documentation or explanation of that performance has been
provided to support the agency’s evaluation rating, which was
equal to Adelphia’s rating. In fact, despite the protester’s
specific complaint of the paucity of the Air Force’s
explanation, the agency has provided no further information of
any kind in support of its evaluation assessment. We find this
inexplicable, given that the RFP provided for a qualitative
assessment of the offerors’ past performance and for an
integrated assessment of the merits of the offerors’ respective
assessments and their evaluated price to determine the “greatest
value.” See RFP at 20. (Southwestern
Bell Telephone Company, B-292476, October 1, 2003) (pdf)
On the record discussed above, we find the agency's
post-protest reevaluation to lack credibility. As discussed
above, the agency has offered no rational support for having
increased AWS's past performance rating. Further, we find this
portion of the agency's reevaluation particularly troubling in
light of the multiple, conflicting numbers that appear in
various post-protest documents regarding the adjusted point
scores to be awarded to Dismas's and AWS's proposals. Similarly,
the agency's summary assertions that Dismas was not prejudiced
by the agency's other procurement errors--including the agency's
failure to consider all of Dismas's proposal information, and
the agency's failure to permit Dismas to respond to adverse past
performance information--are substantially without any
documented, objective analysis.[18] On this record, we decline
to give any material weight to the agency's post-protest
activities and we reject the assertion that Dismas was not
prejudiced. To the contrary, had a proper evaluation been
performed, we believe there is a reasonable possibility that
Dismas's proposal could have been rated higher than AWS's under
a majority of the non-cost/price evaluation factors, including
the most heavily weighted past performance factor. Since the RFP
provided that non-cost/price factors would be “significantly
more important than cost[/price],” we conclude that Dismas has a
substantial chance of receiving the award in the event the
agency properly evaluates Dismas's and AWS's proposals. See
McDonald-Bradley, B‑270126, Feb. 8, 1996, 96-1 CPD ¶ 54 at 3;
see also Statistica, Inc. v. Christopher, 102 F.3d 1577 (Fed.
Cir. 1996). (Dismas Charities,
Inc. , B-292091, June 25, 2003) (pdf)
As indicated above, the evaluation
documentation regarding past performance is sparse and
conclusory. Where an agency fails to document or retain
evaluation materials, it bears the risk that there will be
inadequate supporting rationale in the record for the evaluation
and source selection decision and that we will not conclude that
the agency had a reasonable basis for the decision. Kathpal
Techs., Inc.; Computer & Hi-Tech Mgmt., Inc., B-283137.3 et
al., Dec. 30, 1999, 2000 CPD P: 6 at 12. In addition to the lack
of documentation supporting the past performance evaluation, the
SSA's stated conclusions in the evaluation documentation in this
area were unreasonable or unsupported by the record. (M&S
Farms, Inc., B-290599, September 5, 2002) (pdf)
While an agency is not required to
retain every document or worksheet generated during its
evaluation of proposals, the agency's evaluation must be
sufficiently documented to allow review of the merits of a
protest. KMS Fusion, Inc., B--242529, May 8, 1991, 91-1 CPD para.
447. Where an agency fails to document or retain evaluation
materials, it bears the risk that there may not be adequate
supporting rationale in the record for us to conclude that the
agency had a reasonable basis for the source selection decision.
Southwest Marine, Inc.; American Sys. Eng'g Corp., supra, at 10.
(Northeast
MEP Services, Inc., B-285963.5; B-285963.7; B-285963.8,
January 5, 2001)
Here, we conclude that the
contemporaneous record supports the agency's post-protest
position that the proposals in fact were equivalent under the
technical factor. Specifically, the reference check
questionnaires reflected that the protester and the awardee both
had relevant landscape experience on projects of size and scope
similar to that encompassed by the RFP, and contained uniformly
positive narrative statements concerning their actual
performance. For example, with regard to timeliness of
performance, one of ERG's references stated it was "always
prompt [and] performed within the scope of work," and one
of PLMS's references stated that it was "always
timely--quick response from [named individual]." AR, Tab 6
at 1, 8. With regard to positive and negative performance
factors, the same references stated that ERG's "[p]ersonnel
are very good listeners, dedicated to their projects, and
willing to do and be better in their field of work" and for
PLMS's stated that there were "no negatives[;] very pleased
with contractor." Id. The agency's after-the-fact
scoring is thus wholly consistent with the contemporaneous
record--and thus is entitled to weight in our review--and the
methodology used appears reasonable. (Professional
Landscape Management Services, Inc., B-286612, December 22,
2000)
Under the FSS program, agencies
are not required to conduct a competition before using their
business judgment in determining whether ordering supplies or
services from an FSS vendor represents the best value and meets
the agency's needs at the lowest overall cost. Federal
Acquisition Regulation (FAR) sect. 8.404(a); Amdahl Corp.,
B-281255, Dec. 28, 1998, 98-2 CPD para. 161 at 3. However,
where, as here, an agency conducts a competition, we will review
the agency's actions to ensure that the evaluation and source
selection were reasonable and consistent with the terms of the
solicitation. Computer Prod., Inc., B-284702, May 24, 2000, 2000
CPD para. 95 at 4-5. For such a competition, the agency should
contemporaneously document the basis for its determinations
regarding its needs and the FSS supply or service that meets
those needs at the lowest overall cost in a manner that is
adequate to permit meaningful review; however, in appropriate
circumstances, our Office will consider post-protest evidence
that is a memorialization of the contemporaneous record. Draeger
Safety, Inc., B-285366, B-285366.2, Aug. 23, 2000, 2000 CPD para.
139 at 4, 6; Delta Int'l, Inc., B-284364.2, May 11, 2000, 2000
CPD para. 78 at 4. (Information
Spectrum, Inc., B-285811; B-285811.2, October 17, 2000)
Our review of the record
confirms that the Navy did not adequately document its
evaluation of proposals and that the documentation and further
explanation offered during the course of the protest, including
the two-part hearing that was conducted by our Office, fail to
demonstrate that the evaluation and source selection were
reasonable and supported by the facts. (Future-Tec
Management Systems, Inc.; Computer & Hi-Tech,
B-283793.5; B-283793.6, March 20, 2000)
The contracting officer's Source
Selection/Technical & Cost Evaluation memo provides only the
total technical scores and prices for the competitive range
proposals and percentage comparisons between TMI's technical
score and price and the technical scores and prices of the other
competitive range proposals. The contracting officer's memo
contains no hint as to the basis for the scoring of the
proposals; in fact, there is nothing in the record which
indicates that the contracting officer was ever made aware of
the individually noted strengths and weaknesses in TMI's or
Teltara's proposals. Indeed, the record before us lacks any
evidence that the contracting officer did anything more than
make percentage comparisons among the competitive range proposal
scores and prices in order to determine which offeror should be
awarded the contract. (Teltara
Inc., B-280922, December 4, 1998)
Agency conducted competition on
an unequal basis where the awardee's initial proposal was
substantially in excess of the page limitation stated in the
solicitation, and the agency evaluated that proposal for award
without specifically advising and providing the other offerors
an opportunity to submit proposals without a page limitation.
(Electronic
Design, Inc., B-279662.2; B-279662.3; B-279662.4, August 31,
1998) |