BEST PRACTICES by Gil Harootunian, PhD, Director, ORSP A Dozen Pieces of Tested Advice.
-
Upload
gloria-lawrence -
Category
Documents
-
view
213 -
download
0
Transcript of BEST PRACTICES by Gil Harootunian, PhD, Director, ORSP A Dozen Pieces of Tested Advice.
#1: Program Guidelines
Read (then re-read) the sponsor’s Program Guidelines.
The guidelines clarify explicit ‘eligibility thresholds.’ Research or teaching priorities Funding ceilings, floors, or allowances Standards for evaluation Supplementary requirements or
documents, etc.
#1: Program Guidelines
Entire ‘Program Guidelines’: National Endowment for the Humanities Department of Education National Science Foundation
Outline the critical requirements for complex grants. Create a “Summary Sheet” of critical
points and priorities. “Summary Sheet” is a quick guide
during planning and during submission.
#2: Past Awards
From past awards, you can deduce implicit expectations (vs. explicit expectations in Program Guidelines).
What are they actually funding? You are ‘reading between the lines’ of
the Program Guidelines. Discern Funding Trends.
Department of Education NSF
#2: Past Awards
Other examples National Endowment for the Humanities
National Science Foundation Department of Education
Outline—then consult frequently.
#3: Touch Base w/ ORSP
We will—Translate ‘sponsor-ese’ into plain English.
Share experiential knowledge of what gets funded (and what does not).
Explain both internal and external procedures.
Contact program officers on administrative / budget / technical issues
#3: Touch Base w/ ORSP
We will—Elaborate on sponsor prerogatives like
“positive accountability.”Example: “GRPA”
Elaborate on most common reasons for declension of a grant application.An excessive or poorly justified budget (NSF).
Why? Indicates that project is poorly designed.
(Repeat: We are happy to do budgets!)
#4: Touch Base w/ Program Officer
Example: After you have studied the program guidelines and talked w/ ORSP, contact sponsor’s program officer to see if your idea is competitive.NSF requirementNSF updates
Keep in touch w/ program officer between applications or between pre- and post-award.
#4: Touch Base w/ Program Officer
Cycle back: discuss w/ staff in ORSP the highlights of your discussion w/ sponsor’s program officer.Share and store knowledge for future applications.
Verify program officer’s comments (no human is infallible).
Be careful you do not hear what you want to hear.
#5: Heed the Reviewers
Question: Whom should you heed? Answer: Always heed the
reviewers. Team of experts reviews and—importantly—
responds to your project. Reviewers have more say than anyone else in the
funding of your project. That is why second-time submissions have the
highest funding rate. Most program officers (if asked) will repeat all
this.
#6: Form TeamsDisciplinaryInterdisciplinaryMultidisciplinaryReadersMentors“De facto team” w/ reviewers [who remain anonymous]
#7: Leverage, leverage, leverage
A.k.a., “Bang for the Buck” Example: National Science Foundation’s “EarthCube”
Leverage ≠ Cost ShareCost share = Commitment measured in cash
Leverage = Bringing resources to bear; being resourceful
#8: Network, network, network Read notices in professional journals.
“How I Got My First Grant” tale Ted Stevens’ Amendment
Visit sponsor booths at professional conferences.
If in D.C. area, try to talk w/ program officer(s). If time and resources permit, go to federal
agency workshops. The feds are fast becoming the ‘gold standard’ in
external funding.
#9: Be Resourceful
Use resources. ORSP resources
Sample Budget NarrativesEvaluation guidelinesLetters of support guidelines and samples
Sponsor resourcesNIH “OppNet” ―Basic Behavioral & Social Sciences Opportunity Network
Do not re-invent the wheel.
#10: Disseminate….
Shows you are valuable member
of the field. Advances knowledge in field. Assures sponsor of mileage
(a.k.a., bang for the buck). Does not apply to salary or
proprietary information.
#10: Disseminate….
Examples Articles, conference papers, etc. Fresno State’s ADCC
Faculty who deposit data in digital collections have higher # of citations
NSF: National Science Digital Library
NIH: Pub Med Central NEH: Publicly accessible results
#11: Evaluation—Make It Good
Sponsor ensures it is getting what it wants through on-going evaluation. So important, we have session on this.
Examples: Department of EducationIMLSNEANSF
#11: Evaluation—Make It Good
New standard = positive accountability (vs. old standard of negative accountability)
External vs. internal evaluator? Formative and Summative
Formative = strategy to monitor project as it evolves to provide feedback
Summative = strategy to measure overall effectiveness and extent to which goals met
#12: Align w/ Sponsor’s Priorities
Examples Scalability (a.k.a., Replication).
Does sponsor want a project that can be scaled up regionally or even nationally?
U.S. Department of EducationNational Science Foundation
SustainabilityDoes sponsor want projects that can be
institutionalized?The Council for International Exchange o
f Scholars (a.k.a., Fulbright Program)
#12: Align w/ Sponsor’s Priorities
Examples Community Engagement
Does sponsor want more fluid boundaries between academy and surrounding community?
National Endowment for the Arts Infrastructure
Does sponsor want to strengthen institution’s long-term capacity to support your work?
National Endowment for the Humanities
#12: Align w/ Sponsor’s Priorities
How can you be sure that you are aligning your project’s goals with the sponsor’s goals?
Cycle back: Re-read the program guidelines.
Talk w/ grants managers Talk w/ program officer