This free survey is powered by
QUESTIONPRO.COM
Create a Survey
Surveys
2013
June
P
Program Development Rubric Common Core Spring 2013
Program Development Rubric Common Core Spring 2013
0%
Exit Survey »
This survey was created to be used to score Pivot's Program Development Rubric. You should have received a copy of this rubric in the email with the link to this survey. You should review the rubric as you take this survey. The rubric was designed as a performance metric for Pivot’s program development work. Our plan is to run each of our emerging programs through this rubric at the beginning and end of each year and measure our progress. The rubric is scaled across a four point scale with the following categories (1-4):
1. Conceptual: Few of the essential elements are in place.
2. Emerging: Some of the essential elements are in place and are being piloted.
3. Progressing: Most of the essential elements are in place, have been tested and are now being refined.
4. Replication: All or most of the components have been created, tested and refined.
To make an overall decision about what level your program is at we will (1) take an average of the scores for each of the four program development areas (definition, elements, marketing and scale); then (2) take an average of the four averages you calculated in step one.
Please note that we recognize that some of the areas may not be applicable to your program (e.g. technology tools), just write N/A in the score area and leave that score out of your average.
1. Program Definition: Please rate the CCSS program for the following attributes related to program definition:
1. Conceptual
2. Emerging
3. Progressing
4. Replicating
N/A
*
Does our program have a
clear vision with outcomes
and phases of work? Can we succinctly and coherently describe the key elements of our program?
*
Can we provide
a strong rationale
for why a district should implement the program?
*
Does the program as it is currently defined create a
strong value proposition
for school districts?
*
Do we know what successful implementation looks like (e.g.
performance metrics or milestones
)?
Comments/Evidence/Notes related to your answers for question #1 (Program Definition):
2.
Program Elements/Core Content
: Please rate the CCSS program for the following attributes related to program elements/core content:
1. Conceptual
2. Emerging
3. Progressing
4. Replicating
N/A
*
Have we articulated and defined the content for coaching activities and other
Job Embedded PD activities
?
*
Have we developed a comprehensive set of
PD workshops or training sessions
for our program that can be repurposed for implementation in other districts?
*
Have you developed a viable strategy to build a district’s or school’s
internal organizational structure
to manage the implementation of the program both during the project and after we leave?
*
Has the program created
technology tools
that offer 21st Century solutions to the problem we are trying to address?
*
What
other tools (non technology)
have we create to expedite strategic implementation of the program?
*
Have we identified and developed templates for agreements and policies needed to implement the program?
*
Have we developed an approach to
redesigning a district or school process
that poses a barrier to good teaching and learning, or implementing a new process if one is absent?
Comments/Evidence/Notes related to your answers for question #2 (Program Elements/Core Content):
3. Marketing: Please rate the CCSS program for the following attributes related to marketing/business development:
1. Conceptual
2. Emerging
3. Progressing
4. Replicating
N/A
*
Can we articulate what
services and deliverables
districts would receive and how much they would cost if purchased?
*
Do we have a strong set of
marketing materials
that can be differentiated for different audiences (policy, philanthropy, school districts)?
*
Do we have a
sense of who our customers are
and what funding source they would use to pay for our program?
Comments/Evidence/Notes related to your answers for question #3 (Marketing):
4. Scale Up: Please rate the CCSS program for the following attributes related to scaling up the program (e.g. internal training):
1. Conceptual
2. Emerging
3. Progressing
4. Replicating
N/A
*
Have we identified what
the staffing needs
would be if this program was taken to scale?
*
Would a new staff person be able to implement the program with some basic training? Does the programmatic content come with
sufficient supplemental training material
that it can be delivered with minimal training and oversight/ management?
Comments/Evidence/Notes related to your answers for question #4 (Scale Up/Training):
Loading...
close
Loading...
Close
staticapp1.questionpro.net