SlideShare ist ein Scribd-Unternehmen logo
1 von 38
Data Driven
Instructional Decision Making
A framework
Data –Driven Instruction
Data-driven instruction is characterized by cycles
that provide a feedback loop
in which teachers plan and deliver instruction, assess student
understanding through the collection of data, analyze the data,
and
then pivot instruction based on insights from their analysis.
From: Teachers know best: Making Data Work For Teachers
and Students
Bill & Melinda Gates Foundation
https://s3.amazonaws.com/edtech-production/reports/Gates-
TeachersKnowBest-MakingDataWork.pdf
Data-Driven Decision Making Process Cycle
Data Planning
and
Production
Data Analysis
Developing
an Action
Plan
Monitoring
progress
Measuring
Success
Implementing
the Action
Plan
Data is used
From : Teachers know best: Making Data Work For Teachers
and Students
Bill & Melinda Gates Foundation
https://s3.amazonaws.com/edtech-production/reports/Gates-
TeachersKnowBest-MakingDataWork.pdf
Data –Driven Instruction Feedback Loop
Data Planning
and
Production
Data Analysis
Developing an
Action Plan
Monitoring
progress
Measuring
Success
Implementing
the Action
Plan
Data –Driven Instruction Feedback Loop
Data Planning
and
Production
Data Analysis
Developing an
Action Plan
Monitoring
progress
Measuring
Success
Implementing
the Action
Plan
Instructors need to
facilitate this data –driven
instruction decision loop
in a timely and smooth
fashion
…and on an ongoing basis
• Per student
• Per class
• Per group
Data –Driven Instruction Feedback Loop
Roles Inherent in the Data-Driven Instruction
Decision Making Loop
• Planner
• Data Producer
• Data Analyst
• Monitor
• Reporter
• Data End User
• IT
• Operations and Logistics
Data Planning and Production Questions
• What questions are to be addressed in future data-informed
conversations? Which questions are more important?
• What information (metrics) are needed to answer these
question?
• Is the information available and feasibly attainable?
• Are the necessary technology and resources available?
• How can current non-data based instructional decision making
be
mapped to data-based instructional decision making process?
• What are the costs associated with this endeavor?
• What are the timelines ?
• How and when will the data be collected and stored?
Data Analysis Questions
• What relations exists between the metrics? What patterns do
the data reveal?
• How many levels of the metric are needed to answer the
questions?
• Do the original questions need to be revised or expanded?
• Do the original metrics need to be redefined or expanded?
• What analytical tools are currently available? What tools
need to be designed to support the analysis?
• What method of analysis or evaluation will be used?
• What are the data limitations, strengths, challenges, context?
Monitor Questions
• How are the metrics evolving as the learning and instructional
processes evolve?
• By whom or how are the challenges to the process going to be
detected and addressed?
• How to monitor the improvement processes that were put into
place
as a result of the monitoring?
• What determines whether progress is being made?
Data Driven Instruction Final Project Report
Question: How can Instructors at a distance, educate and
support learners through
the end of the 2020 academic year while maintaining the health
and safety of
learners, staff, and the community?
1) Provide descriptions of how these issues will be addressed in
your fictitious
distance learning setting. Please ensure that you provide a
descriptive overview of
your setting
• Learner Equitable Exposure to Materials
• Learner Accountability
• Learner Personalized Learning Process
• Instructor Work Expectations
2) Determine minimally 2 metrics that could be measured to
monitor and measure
progress and/or challenges related to the 4 issues above
3) Address minimally 3 data planning, analysis, monitoring, and
measurement
questions for each of the 4 issues above
4) Devise and action plan associated with supporting the
fictious Learner
Personalized Learning Process you’ve devised
U.S. Department of Education
September 2013
Five steps for structuring
data‑informed
conversations and action
in education
Wendy Kekahio
Mid-continent Research for
Education and Learning
Myriam Baker
Mid-continent Research for
Education and Learning
Key findings
This guide provides a framework and the tools
and vocabulary needed to support data-informed
conversations and action in education. It walks users
through five key steps in using data for decisionmaking
and strategic action: setting the stage, examining the
data, understanding the findings, developing an action
plan, and monitoring progress and measuring success.
region logo
At Mid-continent Research
for Education and Learning
REL 2013–001
The National Center for Education Evaluation and Regional
Assistance (NCEE) conducts
unbiased large-scale evaluations of education programs and
practices supported by federal
funds; provides research-based technical assistance to educators
and policymakers; and
supports the synthesis and the widespread dissemination of the
results of research and
evaluation throughout the United States.
September 2013
This report was prepared for the Institute of Education Sciences
(IES) under Contract
ED-IES-12-C0010 by Regional Educational Laboratory Pacific
administered by Mid-
continent Research for Education and Learning. The content of
the publication does not
necessarily reflect the views or policies of IES or the U.S.
Department of Education nor
does mention of trade names, commercial products, or
organizations imply endorsement by
the U.S. Government.
This REL report is in the public domain. While permission to
reprint this publication is
not necessary, it should be cited as:
Kekahio, W., & Baker, M. (2013). Five steps for structuring
data-informed conversations and
action in education (REL 2013–001). Washington, DC: U.S.
Department of Education,
Institute of Education Sciences, National Center for Education
Evaluation and Regional
Assistance, Regional Educational Laboratory Pacific. Retrieved
from http://ies.ed.gov/ncee/
edlabs.
This report is available on the Regional Educational Laboratory
website at http://ies.ed.gov/
ncee/edlabs.
http://ies.ed.gov/ncee/edlabs
http://ies.ed.gov/ncee/edlabs
http://ies.ed.gov/ncee/edlabs
http://ies.ed.gov/ncee/edlabs
i
Summary
Using data strategically to guide decisions and actions can have
a positive effect on edu-
cation practices and processes. This facilitation guide shows
education data teams how
to move beyond simply reporting data to applying data to direct
strategic action. Using
guiding questions, suggested activities, and activity forms, this
guide provides education
data teams with a framework and the tools and vocabulary
needed to support an informed
conversation around the data they generate or acquire. The
guide walks data teams
through five key steps in using data for informed
decisionmaking and strategic action:
setting the stage, examining the data, understanding the
findings, developing an action
plan, and monitoring progress and measuring success.
ii
Contents
Summary i
Why this guide? 1
Data and data teams 1
What this guide does 2
Step 1. Setting the stage 2
What is the question? 2
What information is needed to answer the question? 3
Is this information available? 3
Guiding questions 4
Step 2. Examining the data 6
Looking for patterns and making observations 6
Exploring data limitations 6
Guiding questions 7
Step 3. Understanding the findings 7
Choosing a key challenge 7
Brainstorming possible driving factors for strengths and
challenges 7
Guiding questions 10
Step 4. Developing an action plan 10
Setting goals and objectives 10
Developing the plan 12
Guiding questions 12
Step 5. Monitoring progress and measuring success 12
Monitoring progress—how and why? 12
Measuring success—how to know when the goal has been
reached 12
Guiding questions 14
Note Refs-1
References Refs-1
Boxes
1 What is a data team? 1
2 School-based scenario for identifying the “what” question 3
Figure
1 Identifying the driving factors 9
Tables
1 Common types of education data 4
2 SMART goal criteria 10
iii
Templates
1 Setting the stage: recording questions of interest and data
sources 5
2 Examining the data: observation tracking 8
3 Examining the data: strengths and challenges 8
4 Understanding results: driving factor identification 11
5 Developing an action plan: setting SMART goals 13
6 Developing an action plan: organizing the team for action 13
1
Why this guide?
All educators, whatever their role, use data to inform their
decisions. For example, suppose
that math scores in a school have been rising. A data team (box
1) looking at the math
scores for all students in the school might say, “This is great!
These students are clearly
learning a lot.” Or they might say, “Overall, students appear to
be doing well, but are
there some children whose math skills haven’t improved?” In
the first case the team might
respond to the improved math scores by exploring whether the
new after-school math
tutoring program had contributed to the higher scores. In the
second case the team might
look at the scores for each grade and classroom to reveal any
gaps in student gains and
provide more support to lower performing students and their
teachers. Both situations
involve data- informed decisions.
This guide provides grade-, school-, and state-level education
data teams—composed of
teachers, administrators, staff, and other stakeholders—with
steps, sample questions, and
resources for using data more systematically and rigorously in
education decisionmaking.
Data teams can print out the templates in this guide and use
them to direct their own
data-informed conversations.
Data and data teams
Using data to inform education policies and practices is
becoming a greater priority for
educators, spurred by the increased accountability requirements
of the No Child Left
Behind Act of 2001 and by a growing understanding of the
importance of data in edu-
cation decisionmaking. Additionally, states and other entities
are simply collecting more
data. Individual school and institutional datasets are being
shared and, for that reason,
are being structured more consistently. Computer systems and
software allow for quicker
and easier access to data. With the higher stakes associated with
student performance,
growing expectations of education institutions, and increased
awareness of the power of
data, education organizations are expected to be data-
informed—making decisions and
taking action based on the facts presented through data.
Data can enable policymakers to make objective decisions about
education systems,
provide states with information on program effectiveness, and
give teachers and admin-
istrators information on student learning to influence
instruction, programming, and
Box 1. What is a data team?
Data teams are groups of individuals dedicated to data inquiry
and the outcomes it supports.
Data teams address challenges, field questions, and collect
insights from the greater com-
munity of organizations. These teams monitor student and
school performance and efforts to
improve performance within the organization. Data teams also
have primary responsibility for
organizing, analyzing, and coordinating data-informed
responses.
Some data teams are diverse—with representation from school-,
district-, and state-level
personnel, policymakers, researchers, data managers, and family
and community members.
Other data teams comprise individuals with similar roles, such
as teachers of the same grade
or subject or members of a school’s leadership or planning
committee. The composition of
data teams varies by the question of interest, but ideally all
members enrich and strengthen
both the process and the results of data-informed conversations.
This guide provides
education data
teams with steps,
sample questions,
and resources for
using data more
systematically
and rigorously
in education
decisionmaking
2
professional development. Additionally, providing students with
data on their progress
can motivate and engage them in learning (Hamilton et al.,
2009). As educators become
comfortable with the processes described in this guide, they will
be able to convene and
lead their own data teams, supporting colleagues in data inquiry
and building a commit-
ted group of educators accustomed to using data to take action.
Although forming a data
team is not required in order to have a data-informed
conversation, a collaborative team
can identify the resources and personnel needed to collect and
organize data and provide
diverse perspectives on questions, analyses, and actions.
What this guide does
This guide aims to build capacity for using data more
comprehensively and effectively in
the Regional Educational Laboratory (REL) Pacific Region and
elsewhere. It focuses on
helping educators use data that have already been collected—
through the National Center
for Education Statistics, REL Pacific, departments and
ministries of education, school data
systems, analytical and simulation software provided with
textbooks, longitudinal data
systems, and other means.
This guide describes five steps in data-informed conversations
that lead to strategic
decisionmaking and action:
1. Setting the stage. What question is to be addressed in this
data-informed conversation?
What information is needed to answer the question? Is the
information available?
2. Examining the data. What patterns do the data reveal, or what
“snapshot” observa-
tions can be made about the question?
3. Understanding the findings. What are the possible causes for
the patterns?
4. Developing an action plan. How can a data team create an
effective plan for address-
ing the issue?
5. Monitoring progress and measuring success. How can a data
team know whether
progress is being made on the issue?
Each step is described in more detail below.
Step 1. Setting the stage
This step identifies the question to be addressed, the
information needed to answer it, and
the feasibility of accessing the information.
What is the question?
Crafting clear research questions is crucial for enabling data-
informed conversations.
The data team starts by framing the “what” question as simply
as possible. The question
may be very broad to start with, as in “What do the data tell us
about our middle school
students’ math achievement level?” But while it is important to
begin the conversation
by identifying the issue, the question needs to be more specific
(box 2). Narrowing the
question makes identifying and interpreting data clearer and
easier.
Answering the “what” questions can lead to other types of
questions, such as those that
look at relationships between two variables (“Is there a
relationship between middle school
students’ gender and performance on standardized math
achievement tests?”) or those that
Crafting clear
research questions
is crucial
for enabling
data-informed
conversations
3
Box 2. School‑based scenario for identifying the “what”
question
A data team interested in exploring how middle school students
are doing in math might start
with the very general question of “What do the data reveal
about the math achievement of our
middle school students?” Through discussion, the team might
narrow the question to “What
percentage of middle school students are achieving at or above
proficiency on standardized
math achievement tests?” This question identifies a specific
group (middle school students),
measure (standardized math achievement test), and benchmark
or evaluation method (at or
above proficiency). The data team will be guided by this
focused question in examining the
data, understanding the findings, developing an action plan, and
monitoring progress and mea-
suring success.
look at whether one variable influences another (“Does having
an after-school tutoring
program increase middle school math performance on
standardized tests?”).
Keeping long-term goals in mind helps ensure that current and
short-term activities align
with the larger mission. Defining clear short-term objectives
that support long-term goals
also contributes to a continuous data-informed conversation.
Tracking long-term goals
might also require data that are not currently being collected.
Realizing that can initiate a
discussion on future plans for data collection.
What information is needed to answer the question?
After clarifying the question of interest, the data team needs to
identify sources of informa-
tion to answer it. Answering the question posed in box 2 (“What
do the data reveal about
the math achievement of our middle school students?”) requires
access to the results of stan-
dardized math tests for middle school students. Some questions
also require separating data
into smaller groups of interest. For the question on the
relationship between student gender
and standardized math performance, data would need to be
separated by gender.
At this (or any) time, the initial question can be further
narrowed or expanded, but if the
research question changes, the data-use process will need to be
revisited from the beginning.
Is this information available?
It is best to review data sources that have already been analyzed
and reported or that exist
in raw form, such as from departments of education, school-
level records or databases,
and regional educational laboratories—and to bring them all
into the conversation. For
example, some education agencies issue annual reports with
information that has already
been analyzed and reported for entire schools and districts. In
that case further analysis
might not be needed. But if analyses have not already been
done, or have not been com-
pleted in a way that answers the research question, looking for
raw data is the next step. For
the example in box 2 possible sources of standardized math test
results include the Stanford
Achievement Test 10th edition (SAT-10) or state or jurisdiction
benchmark assessments.
Having just one data source might not be enough for
decisionmakers to have confidence in
the results of the analysis. Having multiple sources of
information lessens the likelihood of
ending up without enough data to answer the question.
Keeping long-term
goals in mind helps
ensure that current
and short-term
activities align with
the larger mission
4
Common types of education data include demographic,
perceptual, performance, and
program data (table 1). The type needed depends on the
question.
Data can be collected at the national, state, regional, district,
school, classroom, and student
levels. Data can be gathered at single points in time (cross-
section data) or at multiple
points (longitudinal data) to show how the same student or
groups of students performed
over time. While cross-section data are useful for questions
about what is happening now,
only longitudinal data can provide a comprehensive look at
trends and patterns.
Sometimes the data are insufficient to answer the question. That
in itself is an important
finding. A lack of adequate or reliable data to address important
questions can inform
recommendations for future data collection, even if the
questions cannot yet be answered.
Guiding questions
• What is the question to be addressed in this data-informed
conversation?
• What information is needed to answer the question?
• Is this information available?
Template 1 can be used to record questions and potential data
sources for addressing the
question.
Table 1. Common types of education data
Type of data Description Examples
Demographic These data include descriptors of students,
such as gender, ethnicity, and socioeconomic
status, and descriptors of the organization,
including enrollment and attendance.
• Percentage of students in each ethnic
category.
• Number of students who live within
five miles of the school.
• Average attendance rate for the
school year.
Perceptual These data provide information on how
stakeholders feel or what they observe about
the organization or its activities. They include
stakeholder (student, parent, staff, community
member, graduate) surveys or questionnaires
and observations.
• Parents’ perceptions of school quality.
• Students’ opinions on the importance
of education.
• Community members’ thoughts about
a new school calendar.
Performance These data include information on how
students are performing and on their education
outcomes. The data include information on
types of assessments, grades and grade
point averages, graduation and dropout rates,
mobility rates, suspensions and expulsions,
remediation rates, college acceptance and
attendance rates, and career readiness.
• Percentage of students who scored
proficient or above on the state
standardized assessment.
• Percentage of students who enroll in a
four-year college after high school.
• Number of suspensions in the middle
school over the last school year.
Program These data include descriptive information
on how education and related activities are
conducted within the organization. They include
the textbooks used, the levels of staffing or
professional development at the school, the
schedule of classes, curricular sequences,
instruction strategies, the nature and frequency
of assessments, extracurricular activities, and
even the school setting.
• Teacher-student ratio in the school
district.
• Number of competitive sports offered
in the high school.
• Average number of years of
experience for all elementary school
teachers in the state.
Source: Authors.
Sometimes
the data are
insufficient
to answer the
question; that
in itself is an
important finding
5
Template 1. Setting the stage: recording questions of interest
and data sources
What is the question? What information is needed? What
information is available?
1.
2.
3
4.
5.
Source: Authors.
6
Step 2. Examining the data
Step 2 examines the data for patterns and initial observations
and then explores any data
limitations.
Looking for patterns and making observations
Looking for patterns in the data and making “snapshot”
observations are a first step toward
answering the question. For some questions data from one
period (such as one test or one
school year) may be sufficient. For example, if the question
relates to math proficiency
for the current school year, only one year of data is necessary.
If the question relates to
patterns of middle school math proficiency over time, several
years of data are needed, and
the data team will be looking for patterns across years. Only
after examining the data do
patterns become apparent, whether surprising (such as a steep
increase in average reading
scores in one year), expected (such as average daily student
attendance dropping toward
the end of the school year), or repeated (such as teachers
missing certain items on a certi-
fication exam each year).
During this step it may be possible to begin characterizing
findings from the data as
strengths and challenges. Strengths are results in the data that
indicate a success; challenges
are results that indicate that something is blocking improvement
or higher achievement.
Strengths and challenges can be captured through observations
that arise from examining
the data. These observations should be:
• Specific. Focus on only one item in the data. They are
straightforward and clear
because they involve single sources.
• Factual. Incorporate only the data—no assumptions,
inferences, or biases.
• Related. Begin to answer the question.
For example, while investigating middle school math
achievement, a data team might make
the following observation about a possible success: “Each year
during 2009–12 students showed
a 5 percentage point increase in average achievement scores
between grades 6 and 7.” The fol-
lowing observations might be made about a challenge: “While
student math scores improved
overall, the scores of male students showed a 2 percentage point
decline between 2009 and
2012.” These observations are not opinions but statements of
facts observed in the data.
Exploring data limitations
Before beginning data analysis, it is important to discuss any
limitations in the data, so
that the findings are appropriate. In particular, the data team
should consider whether dif-
ferent sources of data can be compared, as some test data are
not designed for comparison
across grades and years. That is why the increase in math scores
in the previous example
should not be unequivocally considered a success—more
information on the context is
needed to know for sure. Similarly, data that are not the same
(for example, SAT-10 scores
and grade point averages) cannot be combined and analyzed
together. Such data need to
be viewed separately, as they convey different information or
perspectives on the question.
Additionally, while data can inform decisions and provide the
first step in investigating
relationships between programs and outcomes, establishing
causality requires further
Before beginning
data analysis,
it is important
to discuss any
limitations in
the data
7
exploration of the initial observations and findings and rigorous
analysis. Data may not
have been collected or stored in a form that permits
comparisons between groups or that
supports the level of analysis needed to draw robust
conclusions. In that case, the data
team should look into conducting further research and
evaluation studies, if feasible, to
examine any causal relationships of interest.
Guiding questions
• What “snapshot” observations can be made about the question,
or what patterns
are evident?
• Are these observations or patterns related to the original
question? If so, how?
• What strengths or challenges arise from the data?
Template 2 can be used to track observations. Template 3 can be
used to list strengths and
challenges.
Step 3. Understanding the findings
Observations are the basis for identifying the driving factors
that underlie the patterns,
strengths, and challenges shown in the data and for informing
future action plans. For
example, the math results in the previous example could lead a
school principal to say,
“Now that we see there is a challenge in improving math scores
for male students between
grades 6 and 7, let’s figure out why.”
A key step in understanding data results is to begin a discussion
among the data team and
interested colleagues about one or two key challenges.
Understanding why challenges may
be occurring can help address them and might affect the
findings related to the original
question. These driving factors will be the focus of the action
plan for addressing the chal-
lenges and improving results.
Choosing a key challenge
A review of the data often reveals several challenges. The data
team seldom has the time
or resources to explore all the challenges they uncover, so they
will have to focus on just
a few. Some challenges will be more actionable than others and
thus might have a higher
or lower administrative priority. Actionable challenges have
driving factors that educa-
tors can address or influence directly. Other challenges might
be more difficult to address
immediately, such as social factors that influence students at
home (such as lack of super-
vision, low socioeconomic status, single-parent homes). When
possible, data teams should
focus on challenges that are actionable and have a high priority
in the school or district;
however, they should also keep in mind other short- and long-
term objectives, even though
they might not be immediately actionable. These objectives
serve as a broader context for
the team’s ultimate goal.
Brainstorming possible driving factors for strengths and
challenges
One way to identify driving factors for a key challenge (and
thus the extent to which it is
actionable) is to ask a series of “why” questions followed by
“because” responses (figure 1).
The data team can begin to consider possible driving factors
behind the challenge by
understanding the facts and making educated guesses about the
driving factors. As the
While data can
inform decisions
and provide
the first step in
investigating
relationships
between programs
and outcomes,
establishing
causality requires
further exploration
8
Template 2. Examining the data: observation tracking
Data source(s):
Is this observation …
Observations
Related
to the
question?Specific? Factual?
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
Source: Adapted from Nebraska Department of Education
(2012).
Template 3. Examining the data: strengths and challenges
Observed strengths Observed challenges
Example: Over the last three years overall math
achievement scores have improved between grades
6 and 7.
Example: Over the last three years math achievement
scores have declined for male students between
grades 6 and 7.
Source: Adapted from Mid-continent Research for Education
and Learning (2010).
9
Figure 1. Identifying the driving factors
Key challenge: The number of middle school students who meet
or exceed pro�ciency
on math standardized tests has declined during the past two
years.
Challenge (why): Why have test scores declined?
Challenge (why):
Why did they spend
less time on math topics?
Challenge (why):
Why were math topics not
integrated into all subjects?
Challenge (why):
Why were teachers adjusting
to these changes?
Challenge (why):
Why were students learning
different math content?
Potential driving factor (because):
Students spent less time on
math topics each year.
Potential driving factor (because):
Math topics were not integrated
throughout all subjects as in earlier years.
Potential driving factor (because):
Cross-subject team meetings were held
less frequently than in previous years.
Potential driving factor (because):
New standards and curriculum
were introduced with limited
professional development opportunities.
Potential driving factor (because):
Teachers were adjusting to changes
in standards and curriculum.
Potential driving factor (because):
Students were learning different
math content over the past two years.
Source: Authors.
experts in the jurisdiction or school, data team members are the
most familiar with the stu-
dents, issues, and community, and their knowledge forms a
foundation for brainstorming
possible explanations. It is important to stick to the available
facts. Opinions and attitudes
can bias how data are interpreted, and data teams must remain
as objective as possible in
reviewing the data and observations and refrain from drawing
conclusions based on per-
sonal feelings or experiences.
To understand the findings, data teams should revisit any data
limitations. As discussed,
data might provide an initial understanding of the questions but
be unable to support any
causal conclusions. Data teams must recognize any limitations
and uncertainty in the find-
ings as they explore them further.
Once the potential driving factors are identified, data teams
should check that the data
support the ideas. In the example from figure 1 the data team
would need to determine
Once the potential
driving factors
are identified,
data teams should
check that the data
support the ideas
10
whether the observations “Students spend less time on math
topics each year” and “Stu-
dents were learning different math content over the past two
years” are correct by explor-
ing whether adequate data are available on how much time
middle school students have
spent in math classes in the past four to five years. If the
assumptions are supported by the
data, the driving factor can remain. If they are not, the data
team should consider other
potential driving factors that are supported by the data. This
process should be repeated as
many times as needed to identify a strong set of potential
driving factors.
Guiding questions
• What are the possible causes for the observed patterns?
• What school-based factors might have contributed to the
challenges identified in
step 2? Why might these challenges exist?
• Do the data support the identified factors?
Template 4 can be used to identify driving factors.
Step 4. Developing an action plan
Step 4 involves developing an action plan to reduce challenges
and promote successes.
The first part of the plan is to identify the key stakeholders who
will be involved. In the
example, if the jurisdiction or school identifies the math
achievement of male middle
school students as a key challenge, many people will need to be
involved to effectively
address the issue. Without the support of key stakeholders, even
the best plan is unlikely to
succeed. Stakeholders might include school principals,
curriculum and instruction special-
ists, assessment staff, teachers, and parents.
Setting goals and objectives
Careful planning helps data teams take the right actions to
tackle the challenges they
have identified. Action plans typically include short-term
objectives and a long-term goal.
Short-term objectives …
Do all parts separately – separate papers
Part1 – apa style cover and reference needed.
technology innvoations. Those innovations include things like
computer programming (robotics), or computational thinking,
online learning (as we are now doing worldwide!), MOOC's
(Massive open online courses), virtual environments, games and
gamefication, personalized learning systems, and technologies
for students with learning needs. Here is your assignment for
this week. Pick one of the above and tell why you think it has
the most potential for the future, and then how you would/could
use it in your career. (1 and half pages)
Part2- apa style 2 pages cover and reference page needed
submit an essay 2 pages discussing integrating the use of
Bloom's Taxonomy Revised into data-driven instruction
planning for an online course.
https://academicamentoring.com/wp-
content/uploads/2017/01/How-to-use-Blooms-Taxonomy-in-the-
classroom.pdf
readings on the topic of Bloom's Taxonomy, you will be able to
apply what they've read by detecting instances of Bloom's
Taxonomy as students observe an instructor through video
discussing their online course instructional practices. you will
use an online discussion forum to communicate and discuss with
other students what they have detected.
QUestions to address in your essay
1) Which learner level categories of Bloom's Taxonomy Revised
(Remembering, Understanding, Applying, Analyzing,
Evaluating, and Creating) do you believe the instructor is
attempting to address given this learning objective? Please do
not just list the categories but provide specific reasons for your
choices.
2) Given the categories you've identified in #1, do you believe
it is possible for students to successfully reach the learner
category level given the confines of the learning objective? For
example, if you selected Remembering as one of your categories
in #1, does use of the posted resources and the discussion forum
support student attainment of 'Remembering' the taxonomies?
Why or Why not? If not, what revisions or suggestions would
you propose? Please respond relative to each of the categories
you've specified in #1.
3) Given the data-driven instruction decision making learning
process presented in attached pdfs, which stage of this process
do you believe would be most challenging to implement for an
online instructor interested in integrating Bloom's Taxonomy
Revised learning level category objectives for students? Be
specific and choose at least 1 of the Bloom Taxonomy
categories and at least 1 data driven instruction decision making
process step to present your perspective around.
Part3- apa no cover or reference – half pages
https://www.ted.com/talks/peter_norvig_the_100_000_student_c
lassroom
1) Relate ONE category of Bloom's Taxonomy Revised
(Remembering, Understanding, Applying, Analyzing,
Evaluating, Creating) to one of the ideas presented by Dr.
Norvig regarding either his course, instructional practices, or
student observed behaviors. Please be specific in your post
when referencing the video and with which Bloom's Taxonomy
Revised category you are discussing.
2) Discuss your thoughts on how what you've references in #1
above, can be applied in an online course setting.
Data Driven Instructional Decision MakingA framework.docx

Weitere ähnliche Inhalte

Ähnlich wie Data Driven Instructional Decision MakingA framework.docx

07 18-13 webinar - sharnell jackson - using data to personalize learning
07 18-13 webinar - sharnell jackson - using data to personalize learning07 18-13 webinar - sharnell jackson - using data to personalize learning
07 18-13 webinar - sharnell jackson - using data to personalize learningDreamBox Learning
 
How to Use Learning Analytics in Moodle
How to Use Learning Analytics in MoodleHow to Use Learning Analytics in Moodle
How to Use Learning Analytics in MoodleRafael Scapin, Ph.D.
 
ASSESSMENT and EVALUATION.pptx
ASSESSMENT and EVALUATION.pptxASSESSMENT and EVALUATION.pptx
ASSESSMENT and EVALUATION.pptxGIRLYCAMACHO
 
A guide for comprehensive needs assessment
A guide for comprehensive needs assessmentA guide for comprehensive needs assessment
A guide for comprehensive needs assessmentk1hinze
 
A guide for comprehensive needs assessment
A guide for comprehensive needs assessmentA guide for comprehensive needs assessment
A guide for comprehensive needs assessmentk1hinze
 
From theory to practice blending the math classroom and creating a data cultu...
From theory to practice blending the math classroom and creating a data cultu...From theory to practice blending the math classroom and creating a data cultu...
From theory to practice blending the math classroom and creating a data cultu...DreamBox Learning
 
Data teams and cf as (1)
Data teams and cf as (1)Data teams and cf as (1)
Data teams and cf as (1)cbrichards81
 
Using Naviance Data to Drive College & Career Readiness
Using Naviance Data to Drive College & Career ReadinessUsing Naviance Data to Drive College & Career Readiness
Using Naviance Data to Drive College & Career ReadinessNaviance
 
Data/Reform
Data/ReformData/Reform
Data/Reformrgaetano
 
Data Driven Decision Making Presentation
Data Driven Decision Making PresentationData Driven Decision Making Presentation
Data Driven Decision Making PresentationRussell Kunz
 
Swift at isis swis training (2016)
Swift at isis swis training (2016)Swift at isis swis training (2016)
Swift at isis swis training (2016)Katie Conley
 
How data informs decision making 2
How data informs decision making 2How data informs decision making 2
How data informs decision making 2jaccalder
 
School counselors using data
School counselors using dataSchool counselors using data
School counselors using dataaires66
 
Jisc learning analytics service overview Aug 2018
Jisc learning analytics service overview Aug 2018Jisc learning analytics service overview Aug 2018
Jisc learning analytics service overview Aug 2018Paul Bailey
 
Glfes summer institute2013_raleigh_final
Glfes summer institute2013_raleigh_finalGlfes summer institute2013_raleigh_final
Glfes summer institute2013_raleigh_finalTricia Townsend
 
Advanced Foundations and Methods in EL (Lecture 1)
Advanced Foundations and Methods in EL (Lecture 1)Advanced Foundations and Methods in EL (Lecture 1)
Advanced Foundations and Methods in EL (Lecture 1)Sandra Halajian, M.A.
 
Data Visualization and Growth
Data Visualization and GrowthData Visualization and Growth
Data Visualization and GrowthSteven Arnoff
 

Ähnlich wie Data Driven Instructional Decision MakingA framework.docx (20)

07 18-13 webinar - sharnell jackson - using data to personalize learning
07 18-13 webinar - sharnell jackson - using data to personalize learning07 18-13 webinar - sharnell jackson - using data to personalize learning
07 18-13 webinar - sharnell jackson - using data to personalize learning
 
How to Use Learning Analytics in Moodle
How to Use Learning Analytics in MoodleHow to Use Learning Analytics in Moodle
How to Use Learning Analytics in Moodle
 
Data analysis 2011
Data analysis 2011Data analysis 2011
Data analysis 2011
 
ASSESSMENT and EVALUATION.pptx
ASSESSMENT and EVALUATION.pptxASSESSMENT and EVALUATION.pptx
ASSESSMENT and EVALUATION.pptx
 
A guide for comprehensive needs assessment
A guide for comprehensive needs assessmentA guide for comprehensive needs assessment
A guide for comprehensive needs assessment
 
A guide for comprehensive needs assessment
A guide for comprehensive needs assessmentA guide for comprehensive needs assessment
A guide for comprehensive needs assessment
 
From theory to practice blending the math classroom and creating a data cultu...
From theory to practice blending the math classroom and creating a data cultu...From theory to practice blending the math classroom and creating a data cultu...
From theory to practice blending the math classroom and creating a data cultu...
 
Data teams and cf as (1)
Data teams and cf as (1)Data teams and cf as (1)
Data teams and cf as (1)
 
Using Naviance Data to Drive College & Career Readiness
Using Naviance Data to Drive College & Career ReadinessUsing Naviance Data to Drive College & Career Readiness
Using Naviance Data to Drive College & Career Readiness
 
Data/Reform
Data/ReformData/Reform
Data/Reform
 
Data Driven Decision Making Presentation
Data Driven Decision Making PresentationData Driven Decision Making Presentation
Data Driven Decision Making Presentation
 
Swift at isis swis training (2016)
Swift at isis swis training (2016)Swift at isis swis training (2016)
Swift at isis swis training (2016)
 
How data informs decision making 2
How data informs decision making 2How data informs decision making 2
How data informs decision making 2
 
School counselors using data
School counselors using dataSchool counselors using data
School counselors using data
 
Jisc learning analytics service overview Aug 2018
Jisc learning analytics service overview Aug 2018Jisc learning analytics service overview Aug 2018
Jisc learning analytics service overview Aug 2018
 
Glfes summer institute2013_raleigh_final
Glfes summer institute2013_raleigh_finalGlfes summer institute2013_raleigh_final
Glfes summer institute2013_raleigh_final
 
Dai karen
Dai karenDai karen
Dai karen
 
Report 5
Report 5Report 5
Report 5
 
Advanced Foundations and Methods in EL (Lecture 1)
Advanced Foundations and Methods in EL (Lecture 1)Advanced Foundations and Methods in EL (Lecture 1)
Advanced Foundations and Methods in EL (Lecture 1)
 
Data Visualization and Growth
Data Visualization and GrowthData Visualization and Growth
Data Visualization and Growth
 

Mehr von whittemorelucilla

Database reports provide us with the ability to further analyze ou.docx
Database reports provide us with the ability to further analyze ou.docxDatabase reports provide us with the ability to further analyze ou.docx
Database reports provide us with the ability to further analyze ou.docxwhittemorelucilla
 
DataInformationKnowledge1.  Discuss the relationship between.docx
DataInformationKnowledge1.  Discuss the relationship between.docxDataInformationKnowledge1.  Discuss the relationship between.docx
DataInformationKnowledge1.  Discuss the relationship between.docxwhittemorelucilla
 
DataHole 12 Score6757555455555455575775655565656555655656556566643.docx
DataHole 12 Score6757555455555455575775655565656555655656556566643.docxDataHole 12 Score6757555455555455575775655565656555655656556566643.docx
DataHole 12 Score6757555455555455575775655565656555655656556566643.docxwhittemorelucilla
 
DataDestination PalletsTotal CasesCases redCases whiteCases organi.docx
DataDestination PalletsTotal CasesCases redCases whiteCases organi.docxDataDestination PalletsTotal CasesCases redCases whiteCases organi.docx
DataDestination PalletsTotal CasesCases redCases whiteCases organi.docxwhittemorelucilla
 
DataIllinois Tool WorksConsolidated Statement of Income($ in milli.docx
DataIllinois Tool WorksConsolidated Statement of Income($ in milli.docxDataIllinois Tool WorksConsolidated Statement of Income($ in milli.docx
DataIllinois Tool WorksConsolidated Statement of Income($ in milli.docxwhittemorelucilla
 
DataIDSalaryCompa-ratioMidpoint AgePerformance RatingServiceGender.docx
DataIDSalaryCompa-ratioMidpoint AgePerformance RatingServiceGender.docxDataIDSalaryCompa-ratioMidpoint AgePerformance RatingServiceGender.docx
DataIDSalaryCompa-ratioMidpoint AgePerformance RatingServiceGender.docxwhittemorelucilla
 
DataCity1997 Median Price1997 Change1998 Forecast1993-98 Annualize.docx
DataCity1997 Median Price1997 Change1998 Forecast1993-98 Annualize.docxDataCity1997 Median Price1997 Change1998 Forecast1993-98 Annualize.docx
DataCity1997 Median Price1997 Change1998 Forecast1993-98 Annualize.docxwhittemorelucilla
 
DataClientRoom QualityFood QualityService Quality1GPG2GGG3GGG4GPG5.docx
DataClientRoom QualityFood QualityService Quality1GPG2GGG3GGG4GPG5.docxDataClientRoom QualityFood QualityService Quality1GPG2GGG3GGG4GPG5.docx
DataClientRoom QualityFood QualityService Quality1GPG2GGG3GGG4GPG5.docxwhittemorelucilla
 
Database Project CharterBusiness CaseKhalia HartUnive.docx
Database Project CharterBusiness CaseKhalia HartUnive.docxDatabase Project CharterBusiness CaseKhalia HartUnive.docx
Database Project CharterBusiness CaseKhalia HartUnive.docxwhittemorelucilla
 
Databases selected Multiple databases...Full Text (1223 .docx
Databases selected Multiple databases...Full Text (1223  .docxDatabases selected Multiple databases...Full Text (1223  .docx
Databases selected Multiple databases...Full Text (1223 .docxwhittemorelucilla
 
Database SystemsDesign, Implementation, and ManagementCo.docx
Database SystemsDesign, Implementation, and ManagementCo.docxDatabase SystemsDesign, Implementation, and ManagementCo.docx
Database SystemsDesign, Implementation, and ManagementCo.docxwhittemorelucilla
 
DATABASE SYSTEMS DEVELOPMENT & IMPLEMENTATION PLAN1DATABASE SYS.docx
DATABASE SYSTEMS DEVELOPMENT & IMPLEMENTATION PLAN1DATABASE SYS.docxDATABASE SYSTEMS DEVELOPMENT & IMPLEMENTATION PLAN1DATABASE SYS.docx
DATABASE SYSTEMS DEVELOPMENT & IMPLEMENTATION PLAN1DATABASE SYS.docxwhittemorelucilla
 
Database Security Assessment Transcript You are a contracting office.docx
Database Security Assessment Transcript You are a contracting office.docxDatabase Security Assessment Transcript You are a contracting office.docx
Database Security Assessment Transcript You are a contracting office.docxwhittemorelucilla
 
Database Design Mid Term ExamSpring 2020Name ________________.docx
Database Design Mid Term ExamSpring 2020Name ________________.docxDatabase Design Mid Term ExamSpring 2020Name ________________.docx
Database Design Mid Term ExamSpring 2020Name ________________.docxwhittemorelucilla
 
Database Justification MemoCreate a 1-page memo for the .docx
Database Justification MemoCreate a 1-page memo for the .docxDatabase Justification MemoCreate a 1-page memo for the .docx
Database Justification MemoCreate a 1-page memo for the .docxwhittemorelucilla
 
Database Concept Maphttpwikieducator.orgCCNCCCN.docx
Database Concept Maphttpwikieducator.orgCCNCCCN.docxDatabase Concept Maphttpwikieducator.orgCCNCCCN.docx
Database Concept Maphttpwikieducator.orgCCNCCCN.docxwhittemorelucilla
 
Database Dump Script(Details of project in file)Mac1) O.docx
Database Dump Script(Details of project in file)Mac1) O.docxDatabase Dump Script(Details of project in file)Mac1) O.docx
Database Dump Script(Details of project in file)Mac1) O.docxwhittemorelucilla
 
Database Design 1. What is a data model A. method of sto.docx
Database Design 1.  What is a data model A. method of sto.docxDatabase Design 1.  What is a data model A. method of sto.docx
Database Design 1. What is a data model A. method of sto.docxwhittemorelucilla
 
DataAGEGENDERETHNICMAJORSEMHOUSEGPAHRSNEWSPAPTVHRSSLEEPWEIGHTHEIGH.docx
DataAGEGENDERETHNICMAJORSEMHOUSEGPAHRSNEWSPAPTVHRSSLEEPWEIGHTHEIGH.docxDataAGEGENDERETHNICMAJORSEMHOUSEGPAHRSNEWSPAPTVHRSSLEEPWEIGHTHEIGH.docx
DataAGEGENDERETHNICMAJORSEMHOUSEGPAHRSNEWSPAPTVHRSSLEEPWEIGHTHEIGH.docxwhittemorelucilla
 

Mehr von whittemorelucilla (20)

Database reports provide us with the ability to further analyze ou.docx
Database reports provide us with the ability to further analyze ou.docxDatabase reports provide us with the ability to further analyze ou.docx
Database reports provide us with the ability to further analyze ou.docx
 
DataInformationKnowledge1.  Discuss the relationship between.docx
DataInformationKnowledge1.  Discuss the relationship between.docxDataInformationKnowledge1.  Discuss the relationship between.docx
DataInformationKnowledge1.  Discuss the relationship between.docx
 
DataHole 12 Score6757555455555455575775655565656555655656556566643.docx
DataHole 12 Score6757555455555455575775655565656555655656556566643.docxDataHole 12 Score6757555455555455575775655565656555655656556566643.docx
DataHole 12 Score6757555455555455575775655565656555655656556566643.docx
 
DataDestination PalletsTotal CasesCases redCases whiteCases organi.docx
DataDestination PalletsTotal CasesCases redCases whiteCases organi.docxDataDestination PalletsTotal CasesCases redCases whiteCases organi.docx
DataDestination PalletsTotal CasesCases redCases whiteCases organi.docx
 
DataIllinois Tool WorksConsolidated Statement of Income($ in milli.docx
DataIllinois Tool WorksConsolidated Statement of Income($ in milli.docxDataIllinois Tool WorksConsolidated Statement of Income($ in milli.docx
DataIllinois Tool WorksConsolidated Statement of Income($ in milli.docx
 
DataIDSalaryCompa-ratioMidpoint AgePerformance RatingServiceGender.docx
DataIDSalaryCompa-ratioMidpoint AgePerformance RatingServiceGender.docxDataIDSalaryCompa-ratioMidpoint AgePerformance RatingServiceGender.docx
DataIDSalaryCompa-ratioMidpoint AgePerformance RatingServiceGender.docx
 
DataCity1997 Median Price1997 Change1998 Forecast1993-98 Annualize.docx
DataCity1997 Median Price1997 Change1998 Forecast1993-98 Annualize.docxDataCity1997 Median Price1997 Change1998 Forecast1993-98 Annualize.docx
DataCity1997 Median Price1997 Change1998 Forecast1993-98 Annualize.docx
 
DataClientRoom QualityFood QualityService Quality1GPG2GGG3GGG4GPG5.docx
DataClientRoom QualityFood QualityService Quality1GPG2GGG3GGG4GPG5.docxDataClientRoom QualityFood QualityService Quality1GPG2GGG3GGG4GPG5.docx
DataClientRoom QualityFood QualityService Quality1GPG2GGG3GGG4GPG5.docx
 
Database Project CharterBusiness CaseKhalia HartUnive.docx
Database Project CharterBusiness CaseKhalia HartUnive.docxDatabase Project CharterBusiness CaseKhalia HartUnive.docx
Database Project CharterBusiness CaseKhalia HartUnive.docx
 
Databases selected Multiple databases...Full Text (1223 .docx
Databases selected Multiple databases...Full Text (1223  .docxDatabases selected Multiple databases...Full Text (1223  .docx
Databases selected Multiple databases...Full Text (1223 .docx
 
Database SystemsDesign, Implementation, and ManagementCo.docx
Database SystemsDesign, Implementation, and ManagementCo.docxDatabase SystemsDesign, Implementation, and ManagementCo.docx
Database SystemsDesign, Implementation, and ManagementCo.docx
 
DATABASE SYSTEMS DEVELOPMENT & IMPLEMENTATION PLAN1DATABASE SYS.docx
DATABASE SYSTEMS DEVELOPMENT & IMPLEMENTATION PLAN1DATABASE SYS.docxDATABASE SYSTEMS DEVELOPMENT & IMPLEMENTATION PLAN1DATABASE SYS.docx
DATABASE SYSTEMS DEVELOPMENT & IMPLEMENTATION PLAN1DATABASE SYS.docx
 
Database Security Assessment Transcript You are a contracting office.docx
Database Security Assessment Transcript You are a contracting office.docxDatabase Security Assessment Transcript You are a contracting office.docx
Database Security Assessment Transcript You are a contracting office.docx
 
Data.docx
Data.docxData.docx
Data.docx
 
Database Design Mid Term ExamSpring 2020Name ________________.docx
Database Design Mid Term ExamSpring 2020Name ________________.docxDatabase Design Mid Term ExamSpring 2020Name ________________.docx
Database Design Mid Term ExamSpring 2020Name ________________.docx
 
Database Justification MemoCreate a 1-page memo for the .docx
Database Justification MemoCreate a 1-page memo for the .docxDatabase Justification MemoCreate a 1-page memo for the .docx
Database Justification MemoCreate a 1-page memo for the .docx
 
Database Concept Maphttpwikieducator.orgCCNCCCN.docx
Database Concept Maphttpwikieducator.orgCCNCCCN.docxDatabase Concept Maphttpwikieducator.orgCCNCCCN.docx
Database Concept Maphttpwikieducator.orgCCNCCCN.docx
 
Database Dump Script(Details of project in file)Mac1) O.docx
Database Dump Script(Details of project in file)Mac1) O.docxDatabase Dump Script(Details of project in file)Mac1) O.docx
Database Dump Script(Details of project in file)Mac1) O.docx
 
Database Design 1. What is a data model A. method of sto.docx
Database Design 1.  What is a data model A. method of sto.docxDatabase Design 1.  What is a data model A. method of sto.docx
Database Design 1. What is a data model A. method of sto.docx
 
DataAGEGENDERETHNICMAJORSEMHOUSEGPAHRSNEWSPAPTVHRSSLEEPWEIGHTHEIGH.docx
DataAGEGENDERETHNICMAJORSEMHOUSEGPAHRSNEWSPAPTVHRSSLEEPWEIGHTHEIGH.docxDataAGEGENDERETHNICMAJORSEMHOUSEGPAHRSNEWSPAPTVHRSSLEEPWEIGHTHEIGH.docx
DataAGEGENDERETHNICMAJORSEMHOUSEGPAHRSNEWSPAPTVHRSSLEEPWEIGHTHEIGH.docx
 

Kürzlich hochgeladen

A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformChameera Dedduwage
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room servicediscovermytutordmt
 
social pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajansocial pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajanpragatimahajan3
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...christianmathematics
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfAdmir Softic
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsTechSoup
 
APM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAPM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAssociation for Project Management
 
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdfBASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdfSoniaTolstoy
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13Steve Thomason
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDThiyagu K
 
Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Disha Kariya
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphThiyagu K
 
Disha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfDisha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfchloefrazer622
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxVishalSingh1417
 
Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3JemimahLaneBuaron
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfAyushMahapatra5
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfsanyamsingh5019
 

Kürzlich hochgeladen (20)

A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy Reform
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room service
 
social pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajansocial pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajan
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
APM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAPM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across Sectors
 
Advance Mobile Application Development class 07
Advance Mobile Application Development class 07Advance Mobile Application Development class 07
Advance Mobile Application Development class 07
 
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdfBASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SD
 
Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot Graph
 
Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"
 
Disha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfDisha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdf
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 
Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3
 
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdf
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdf
 

Data Driven Instructional Decision MakingA framework.docx

  • 1. Data Driven Instructional Decision Making A framework Data –Driven Instruction Data-driven instruction is characterized by cycles that provide a feedback loop in which teachers plan and deliver instruction, assess student understanding through the collection of data, analyze the data, and then pivot instruction based on insights from their analysis. From: Teachers know best: Making Data Work For Teachers and Students Bill & Melinda Gates Foundation https://s3.amazonaws.com/edtech-production/reports/Gates- TeachersKnowBest-MakingDataWork.pdf Data-Driven Decision Making Process Cycle Data Planning and Production
  • 2. Data Analysis Developing an Action Plan Monitoring progress Measuring Success Implementing the Action Plan Data is used From : Teachers know best: Making Data Work For Teachers and Students Bill & Melinda Gates Foundation https://s3.amazonaws.com/edtech-production/reports/Gates- TeachersKnowBest-MakingDataWork.pdf Data –Driven Instruction Feedback Loop Data Planning
  • 3. and Production Data Analysis Developing an Action Plan Monitoring progress Measuring Success Implementing the Action Plan Data –Driven Instruction Feedback Loop Data Planning and Production Data Analysis Developing an Action Plan Monitoring progress
  • 4. Measuring Success Implementing the Action Plan Instructors need to facilitate this data –driven instruction decision loop in a timely and smooth fashion …and on an ongoing basis • Per student • Per class • Per group Data –Driven Instruction Feedback Loop Roles Inherent in the Data-Driven Instruction Decision Making Loop • Planner • Data Producer • Data Analyst • Monitor • Reporter
  • 5. • Data End User • IT • Operations and Logistics Data Planning and Production Questions • What questions are to be addressed in future data-informed conversations? Which questions are more important? • What information (metrics) are needed to answer these question? • Is the information available and feasibly attainable? • Are the necessary technology and resources available? • How can current non-data based instructional decision making be mapped to data-based instructional decision making process? • What are the costs associated with this endeavor? • What are the timelines ? • How and when will the data be collected and stored? Data Analysis Questions • What relations exists between the metrics? What patterns do the data reveal? • How many levels of the metric are needed to answer the questions?
  • 6. • Do the original questions need to be revised or expanded? • Do the original metrics need to be redefined or expanded? • What analytical tools are currently available? What tools need to be designed to support the analysis? • What method of analysis or evaluation will be used? • What are the data limitations, strengths, challenges, context? Monitor Questions • How are the metrics evolving as the learning and instructional processes evolve? • By whom or how are the challenges to the process going to be detected and addressed? • How to monitor the improvement processes that were put into place as a result of the monitoring? • What determines whether progress is being made? Data Driven Instruction Final Project Report Question: How can Instructors at a distance, educate and support learners through the end of the 2020 academic year while maintaining the health and safety of learners, staff, and the community? 1) Provide descriptions of how these issues will be addressed in your fictitious distance learning setting. Please ensure that you provide a
  • 7. descriptive overview of your setting • Learner Equitable Exposure to Materials • Learner Accountability • Learner Personalized Learning Process • Instructor Work Expectations 2) Determine minimally 2 metrics that could be measured to monitor and measure progress and/or challenges related to the 4 issues above 3) Address minimally 3 data planning, analysis, monitoring, and measurement questions for each of the 4 issues above 4) Devise and action plan associated with supporting the fictious Learner Personalized Learning Process you’ve devised U.S. Department of Education September 2013 Five steps for structuring data‑informed conversations and action in education Wendy Kekahio Mid-continent Research for Education and Learning
  • 8. Myriam Baker Mid-continent Research for Education and Learning Key findings This guide provides a framework and the tools and vocabulary needed to support data-informed conversations and action in education. It walks users through five key steps in using data for decisionmaking and strategic action: setting the stage, examining the data, understanding the findings, developing an action plan, and monitoring progress and measuring success. region logo At Mid-continent Research for Education and Learning REL 2013–001 The National Center for Education Evaluation and Regional Assistance (NCEE) conducts unbiased large-scale evaluations of education programs and practices supported by federal funds; provides research-based technical assistance to educators
  • 9. and policymakers; and supports the synthesis and the widespread dissemination of the results of research and evaluation throughout the United States. September 2013 This report was prepared for the Institute of Education Sciences (IES) under Contract ED-IES-12-C0010 by Regional Educational Laboratory Pacific administered by Mid- continent Research for Education and Learning. The content of the publication does not necessarily reflect the views or policies of IES or the U.S. Department of Education nor does mention of trade names, commercial products, or organizations imply endorsement by the U.S. Government. This REL report is in the public domain. While permission to reprint this publication is not necessary, it should be cited as: Kekahio, W., & Baker, M. (2013). Five steps for structuring data-informed conversations and action in education (REL 2013–001). Washington, DC: U.S. Department of Education, Institute of Education Sciences, National Center for Education Evaluation and Regional Assistance, Regional Educational Laboratory Pacific. Retrieved from http://ies.ed.gov/ncee/ edlabs. This report is available on the Regional Educational Laboratory website at http://ies.ed.gov/ ncee/edlabs.
  • 10. http://ies.ed.gov/ncee/edlabs http://ies.ed.gov/ncee/edlabs http://ies.ed.gov/ncee/edlabs http://ies.ed.gov/ncee/edlabs i Summary Using data strategically to guide decisions and actions can have a positive effect on edu- cation practices and processes. This facilitation guide shows education data teams how to move beyond simply reporting data to applying data to direct strategic action. Using guiding questions, suggested activities, and activity forms, this guide provides education data teams with a framework and the tools and vocabulary needed to support an informed conversation around the data they generate or acquire. The guide walks data teams through five key steps in using data for informed decisionmaking and strategic action: setting the stage, examining the data, understanding the findings, developing an action plan, and monitoring progress and measuring success. ii Contents Summary i
  • 11. Why this guide? 1 Data and data teams 1 What this guide does 2 Step 1. Setting the stage 2 What is the question? 2 What information is needed to answer the question? 3 Is this information available? 3 Guiding questions 4 Step 2. Examining the data 6 Looking for patterns and making observations 6 Exploring data limitations 6 Guiding questions 7 Step 3. Understanding the findings 7 Choosing a key challenge 7 Brainstorming possible driving factors for strengths and challenges 7 Guiding questions 10 Step 4. Developing an action plan 10 Setting goals and objectives 10 Developing the plan 12 Guiding questions 12 Step 5. Monitoring progress and measuring success 12 Monitoring progress—how and why? 12 Measuring success—how to know when the goal has been reached 12 Guiding questions 14 Note Refs-1 References Refs-1
  • 12. Boxes 1 What is a data team? 1 2 School-based scenario for identifying the “what” question 3 Figure 1 Identifying the driving factors 9 Tables 1 Common types of education data 4 2 SMART goal criteria 10 iii Templates 1 Setting the stage: recording questions of interest and data sources 5 2 Examining the data: observation tracking 8 3 Examining the data: strengths and challenges 8 4 Understanding results: driving factor identification 11 5 Developing an action plan: setting SMART goals 13 6 Developing an action plan: organizing the team for action 13 1 Why this guide? All educators, whatever their role, use data to inform their decisions. For example, suppose that math scores in a school have been rising. A data team (box 1) looking at the math scores for all students in the school might say, “This is great!
  • 13. These students are clearly learning a lot.” Or they might say, “Overall, students appear to be doing well, but are there some children whose math skills haven’t improved?” In the first case the team might respond to the improved math scores by exploring whether the new after-school math tutoring program had contributed to the higher scores. In the second case the team might look at the scores for each grade and classroom to reveal any gaps in student gains and provide more support to lower performing students and their teachers. Both situations involve data- informed decisions. This guide provides grade-, school-, and state-level education data teams—composed of teachers, administrators, staff, and other stakeholders—with steps, sample questions, and resources for using data more systematically and rigorously in education decisionmaking. Data teams can print out the templates in this guide and use them to direct their own data-informed conversations. Data and data teams Using data to inform education policies and practices is becoming a greater priority for educators, spurred by the increased accountability requirements of the No Child Left Behind Act of 2001 and by a growing understanding of the importance of data in edu- cation decisionmaking. Additionally, states and other entities are simply collecting more data. Individual school and institutional datasets are being
  • 14. shared and, for that reason, are being structured more consistently. Computer systems and software allow for quicker and easier access to data. With the higher stakes associated with student performance, growing expectations of education institutions, and increased awareness of the power of data, education organizations are expected to be data- informed—making decisions and taking action based on the facts presented through data. Data can enable policymakers to make objective decisions about education systems, provide states with information on program effectiveness, and give teachers and admin- istrators information on student learning to influence instruction, programming, and Box 1. What is a data team? Data teams are groups of individuals dedicated to data inquiry and the outcomes it supports. Data teams address challenges, field questions, and collect insights from the greater com- munity of organizations. These teams monitor student and school performance and efforts to improve performance within the organization. Data teams also have primary responsibility for organizing, analyzing, and coordinating data-informed responses. Some data teams are diverse—with representation from school-,
  • 15. district-, and state-level personnel, policymakers, researchers, data managers, and family and community members. Other data teams comprise individuals with similar roles, such as teachers of the same grade or subject or members of a school’s leadership or planning committee. The composition of data teams varies by the question of interest, but ideally all members enrich and strengthen both the process and the results of data-informed conversations. This guide provides education data teams with steps, sample questions, and resources for using data more systematically and rigorously in education decisionmaking 2 professional development. Additionally, providing students with data on their progress can motivate and engage them in learning (Hamilton et al., 2009). As educators become comfortable with the processes described in this guide, they will
  • 16. be able to convene and lead their own data teams, supporting colleagues in data inquiry and building a commit- ted group of educators accustomed to using data to take action. Although forming a data team is not required in order to have a data-informed conversation, a collaborative team can identify the resources and personnel needed to collect and organize data and provide diverse perspectives on questions, analyses, and actions. What this guide does This guide aims to build capacity for using data more comprehensively and effectively in the Regional Educational Laboratory (REL) Pacific Region and elsewhere. It focuses on helping educators use data that have already been collected— through the National Center for Education Statistics, REL Pacific, departments and ministries of education, school data systems, analytical and simulation software provided with textbooks, longitudinal data systems, and other means. This guide describes five steps in data-informed conversations that lead to strategic decisionmaking and action: 1. Setting the stage. What question is to be addressed in this data-informed conversation? What information is needed to answer the question? Is the information available? 2. Examining the data. What patterns do the data reveal, or what “snapshot” observa-
  • 17. tions can be made about the question? 3. Understanding the findings. What are the possible causes for the patterns? 4. Developing an action plan. How can a data team create an effective plan for address- ing the issue? 5. Monitoring progress and measuring success. How can a data team know whether progress is being made on the issue? Each step is described in more detail below. Step 1. Setting the stage This step identifies the question to be addressed, the information needed to answer it, and the feasibility of accessing the information. What is the question? Crafting clear research questions is crucial for enabling data- informed conversations. The data team starts by framing the “what” question as simply as possible. The question may be very broad to start with, as in “What do the data tell us about our middle school students’ math achievement level?” But while it is important to begin the conversation by identifying the issue, the question needs to be more specific (box 2). Narrowing the question makes identifying and interpreting data clearer and easier.
  • 18. Answering the “what” questions can lead to other types of questions, such as those that look at relationships between two variables (“Is there a relationship between middle school students’ gender and performance on standardized math achievement tests?”) or those that Crafting clear research questions is crucial for enabling data-informed conversations 3 Box 2. School‑based scenario for identifying the “what” question A data team interested in exploring how middle school students are doing in math might start with the very general question of “What do the data reveal about the math achievement of our middle school students?” Through discussion, the team might narrow the question to “What percentage of middle school students are achieving at or above proficiency on standardized math achievement tests?” This question identifies a specific group (middle school students),
  • 19. measure (standardized math achievement test), and benchmark or evaluation method (at or above proficiency). The data team will be guided by this focused question in examining the data, understanding the findings, developing an action plan, and monitoring progress and mea- suring success. look at whether one variable influences another (“Does having an after-school tutoring program increase middle school math performance on standardized tests?”). Keeping long-term goals in mind helps ensure that current and short-term activities align with the larger mission. Defining clear short-term objectives that support long-term goals also contributes to a continuous data-informed conversation. Tracking long-term goals might also require data that are not currently being collected. Realizing that can initiate a discussion on future plans for data collection. What information is needed to answer the question? After clarifying the question of interest, the data team needs to identify sources of informa- tion to answer it. Answering the question posed in box 2 (“What do the data reveal about the math achievement of our middle school students?”) requires access to the results of stan- dardized math tests for middle school students. Some questions also require separating data
  • 20. into smaller groups of interest. For the question on the relationship between student gender and standardized math performance, data would need to be separated by gender. At this (or any) time, the initial question can be further narrowed or expanded, but if the research question changes, the data-use process will need to be revisited from the beginning. Is this information available? It is best to review data sources that have already been analyzed and reported or that exist in raw form, such as from departments of education, school- level records or databases, and regional educational laboratories—and to bring them all into the conversation. For example, some education agencies issue annual reports with information that has already been analyzed and reported for entire schools and districts. In that case further analysis might not be needed. But if analyses have not already been done, or have not been com- pleted in a way that answers the research question, looking for raw data is the next step. For the example in box 2 possible sources of standardized math test results include the Stanford Achievement Test 10th edition (SAT-10) or state or jurisdiction benchmark assessments. Having just one data source might not be enough for decisionmakers to have confidence in the results of the analysis. Having multiple sources of information lessens the likelihood of ending up without enough data to answer the question.
  • 21. Keeping long-term goals in mind helps ensure that current and short-term activities align with the larger mission 4 Common types of education data include demographic, perceptual, performance, and program data (table 1). The type needed depends on the question. Data can be collected at the national, state, regional, district, school, classroom, and student levels. Data can be gathered at single points in time (cross- section data) or at multiple points (longitudinal data) to show how the same student or groups of students performed over time. While cross-section data are useful for questions about what is happening now, only longitudinal data can provide a comprehensive look at trends and patterns. Sometimes the data are insufficient to answer the question. That in itself is an important finding. A lack of adequate or reliable data to address important questions can inform recommendations for future data collection, even if the questions cannot yet be answered. Guiding questions
  • 22. • What is the question to be addressed in this data-informed conversation? • What information is needed to answer the question? • Is this information available? Template 1 can be used to record questions and potential data sources for addressing the question. Table 1. Common types of education data Type of data Description Examples Demographic These data include descriptors of students, such as gender, ethnicity, and socioeconomic status, and descriptors of the organization, including enrollment and attendance. • Percentage of students in each ethnic category. • Number of students who live within five miles of the school. • Average attendance rate for the school year. Perceptual These data provide information on how stakeholders feel or what they observe about the organization or its activities. They include stakeholder (student, parent, staff, community member, graduate) surveys or questionnaires and observations. • Parents’ perceptions of school quality. • Students’ opinions on the importance
  • 23. of education. • Community members’ thoughts about a new school calendar. Performance These data include information on how students are performing and on their education outcomes. The data include information on types of assessments, grades and grade point averages, graduation and dropout rates, mobility rates, suspensions and expulsions, remediation rates, college acceptance and attendance rates, and career readiness. • Percentage of students who scored proficient or above on the state standardized assessment. • Percentage of students who enroll in a four-year college after high school. • Number of suspensions in the middle school over the last school year. Program These data include descriptive information on how education and related activities are conducted within the organization. They include the textbooks used, the levels of staffing or professional development at the school, the schedule of classes, curricular sequences, instruction strategies, the nature and frequency of assessments, extracurricular activities, and even the school setting. • Teacher-student ratio in the school
  • 24. district. • Number of competitive sports offered in the high school. • Average number of years of experience for all elementary school teachers in the state. Source: Authors. Sometimes the data are insufficient to answer the question; that in itself is an important finding 5 Template 1. Setting the stage: recording questions of interest and data sources What is the question? What information is needed? What information is available? 1. 2. 3 4.
  • 25. 5. Source: Authors. 6 Step 2. Examining the data Step 2 examines the data for patterns and initial observations and then explores any data limitations. Looking for patterns and making observations Looking for patterns in the data and making “snapshot” observations are a first step toward answering the question. For some questions data from one period (such as one test or one school year) may be sufficient. For example, if the question relates to math proficiency for the current school year, only one year of data is necessary. If the question relates to patterns of middle school math proficiency over time, several years of data are needed, and the data team will be looking for patterns across years. Only after examining the data do patterns become apparent, whether surprising (such as a steep increase in average reading scores in one year), expected (such as average daily student attendance dropping toward the end of the school year), or repeated (such as teachers missing certain items on a certi- fication exam each year).
  • 26. During this step it may be possible to begin characterizing findings from the data as strengths and challenges. Strengths are results in the data that indicate a success; challenges are results that indicate that something is blocking improvement or higher achievement. Strengths and challenges can be captured through observations that arise from examining the data. These observations should be: • Specific. Focus on only one item in the data. They are straightforward and clear because they involve single sources. • Factual. Incorporate only the data—no assumptions, inferences, or biases. • Related. Begin to answer the question. For example, while investigating middle school math achievement, a data team might make the following observation about a possible success: “Each year during 2009–12 students showed a 5 percentage point increase in average achievement scores between grades 6 and 7.” The fol- lowing observations might be made about a challenge: “While student math scores improved overall, the scores of male students showed a 2 percentage point decline between 2009 and 2012.” These observations are not opinions but statements of facts observed in the data. Exploring data limitations Before beginning data analysis, it is important to discuss any
  • 27. limitations in the data, so that the findings are appropriate. In particular, the data team should consider whether dif- ferent sources of data can be compared, as some test data are not designed for comparison across grades and years. That is why the increase in math scores in the previous example should not be unequivocally considered a success—more information on the context is needed to know for sure. Similarly, data that are not the same (for example, SAT-10 scores and grade point averages) cannot be combined and analyzed together. Such data need to be viewed separately, as they convey different information or perspectives on the question. Additionally, while data can inform decisions and provide the first step in investigating relationships between programs and outcomes, establishing causality requires further Before beginning data analysis, it is important to discuss any limitations in the data 7 exploration of the initial observations and findings and rigorous analysis. Data may not have been collected or stored in a form that permits comparisons between groups or that
  • 28. supports the level of analysis needed to draw robust conclusions. In that case, the data team should look into conducting further research and evaluation studies, if feasible, to examine any causal relationships of interest. Guiding questions • What “snapshot” observations can be made about the question, or what patterns are evident? • Are these observations or patterns related to the original question? If so, how? • What strengths or challenges arise from the data? Template 2 can be used to track observations. Template 3 can be used to list strengths and challenges. Step 3. Understanding the findings Observations are the basis for identifying the driving factors that underlie the patterns, strengths, and challenges shown in the data and for informing future action plans. For example, the math results in the previous example could lead a school principal to say, “Now that we see there is a challenge in improving math scores for male students between grades 6 and 7, let’s figure out why.” A key step in understanding data results is to begin a discussion among the data team and interested colleagues about one or two key challenges. Understanding why challenges may be occurring can help address them and might affect the
  • 29. findings related to the original question. These driving factors will be the focus of the action plan for addressing the chal- lenges and improving results. Choosing a key challenge A review of the data often reveals several challenges. The data team seldom has the time or resources to explore all the challenges they uncover, so they will have to focus on just a few. Some challenges will be more actionable than others and thus might have a higher or lower administrative priority. Actionable challenges have driving factors that educa- tors can address or influence directly. Other challenges might be more difficult to address immediately, such as social factors that influence students at home (such as lack of super- vision, low socioeconomic status, single-parent homes). When possible, data teams should focus on challenges that are actionable and have a high priority in the school or district; however, they should also keep in mind other short- and long- term objectives, even though they might not be immediately actionable. These objectives serve as a broader context for the team’s ultimate goal. Brainstorming possible driving factors for strengths and challenges One way to identify driving factors for a key challenge (and thus the extent to which it is actionable) is to ask a series of “why” questions followed by “because” responses (figure 1).
  • 30. The data team can begin to consider possible driving factors behind the challenge by understanding the facts and making educated guesses about the driving factors. As the While data can inform decisions and provide the first step in investigating relationships between programs and outcomes, establishing causality requires further exploration 8 Template 2. Examining the data: observation tracking Data source(s): Is this observation … Observations Related to the question?Specific? Factual? 1.
  • 31. 2. 3. 4. 5. 6. 7. 8. 9. 10. Source: Adapted from Nebraska Department of Education (2012). Template 3. Examining the data: strengths and challenges Observed strengths Observed challenges Example: Over the last three years overall math achievement scores have improved between grades 6 and 7. Example: Over the last three years math achievement scores have declined for male students between grades 6 and 7. Source: Adapted from Mid-continent Research for Education and Learning (2010).
  • 32. 9 Figure 1. Identifying the driving factors Key challenge: The number of middle school students who meet or exceed pro�ciency on math standardized tests has declined during the past two years. Challenge (why): Why have test scores declined? Challenge (why): Why did they spend less time on math topics? Challenge (why): Why were math topics not integrated into all subjects? Challenge (why): Why were teachers adjusting to these changes? Challenge (why): Why were students learning different math content? Potential driving factor (because): Students spent less time on
  • 33. math topics each year. Potential driving factor (because): Math topics were not integrated throughout all subjects as in earlier years. Potential driving factor (because): Cross-subject team meetings were held less frequently than in previous years. Potential driving factor (because): New standards and curriculum were introduced with limited professional development opportunities. Potential driving factor (because): Teachers were adjusting to changes in standards and curriculum. Potential driving factor (because): Students were learning different math content over the past two years. Source: Authors. experts in the jurisdiction or school, data team members are the most familiar with the stu- dents, issues, and community, and their knowledge forms a foundation for brainstorming possible explanations. It is important to stick to the available facts. Opinions and attitudes can bias how data are interpreted, and data teams must remain
  • 34. as objective as possible in reviewing the data and observations and refrain from drawing conclusions based on per- sonal feelings or experiences. To understand the findings, data teams should revisit any data limitations. As discussed, data might provide an initial understanding of the questions but be unable to support any causal conclusions. Data teams must recognize any limitations and uncertainty in the find- ings as they explore them further. Once the potential driving factors are identified, data teams should check that the data support the ideas. In the example from figure 1 the data team would need to determine Once the potential driving factors are identified, data teams should check that the data support the ideas 10 whether the observations “Students spend less time on math topics each year” and “Stu- dents were learning different math content over the past two years” are correct by explor- ing whether adequate data are available on how much time middle school students have spent in math classes in the past four to five years. If the
  • 35. assumptions are supported by the data, the driving factor can remain. If they are not, the data team should consider other potential driving factors that are supported by the data. This process should be repeated as many times as needed to identify a strong set of potential driving factors. Guiding questions • What are the possible causes for the observed patterns? • What school-based factors might have contributed to the challenges identified in step 2? Why might these challenges exist? • Do the data support the identified factors? Template 4 can be used to identify driving factors. Step 4. Developing an action plan Step 4 involves developing an action plan to reduce challenges and promote successes. The first part of the plan is to identify the key stakeholders who will be involved. In the example, if the jurisdiction or school identifies the math achievement of male middle school students as a key challenge, many people will need to be involved to effectively address the issue. Without the support of key stakeholders, even the best plan is unlikely to succeed. Stakeholders might include school principals, curriculum and instruction special- ists, assessment staff, teachers, and parents. Setting goals and objectives
  • 36. Careful planning helps data teams take the right actions to tackle the challenges they have identified. Action plans typically include short-term objectives and a long-term goal. Short-term objectives … Do all parts separately – separate papers Part1 – apa style cover and reference needed. technology innvoations. Those innovations include things like computer programming (robotics), or computational thinking, online learning (as we are now doing worldwide!), MOOC's (Massive open online courses), virtual environments, games and gamefication, personalized learning systems, and technologies for students with learning needs. Here is your assignment for this week. Pick one of the above and tell why you think it has the most potential for the future, and then how you would/could use it in your career. (1 and half pages) Part2- apa style 2 pages cover and reference page needed submit an essay 2 pages discussing integrating the use of Bloom's Taxonomy Revised into data-driven instruction planning for an online course. https://academicamentoring.com/wp- content/uploads/2017/01/How-to-use-Blooms-Taxonomy-in-the- classroom.pdf readings on the topic of Bloom's Taxonomy, you will be able to apply what they've read by detecting instances of Bloom's Taxonomy as students observe an instructor through video discussing their online course instructional practices. you will use an online discussion forum to communicate and discuss with other students what they have detected. QUestions to address in your essay 1) Which learner level categories of Bloom's Taxonomy Revised (Remembering, Understanding, Applying, Analyzing, Evaluating, and Creating) do you believe the instructor is
  • 37. attempting to address given this learning objective? Please do not just list the categories but provide specific reasons for your choices. 2) Given the categories you've identified in #1, do you believe it is possible for students to successfully reach the learner category level given the confines of the learning objective? For example, if you selected Remembering as one of your categories in #1, does use of the posted resources and the discussion forum support student attainment of 'Remembering' the taxonomies? Why or Why not? If not, what revisions or suggestions would you propose? Please respond relative to each of the categories you've specified in #1. 3) Given the data-driven instruction decision making learning process presented in attached pdfs, which stage of this process do you believe would be most challenging to implement for an online instructor interested in integrating Bloom's Taxonomy Revised learning level category objectives for students? Be specific and choose at least 1 of the Bloom Taxonomy categories and at least 1 data driven instruction decision making process step to present your perspective around. Part3- apa no cover or reference – half pages https://www.ted.com/talks/peter_norvig_the_100_000_student_c lassroom 1) Relate ONE category of Bloom's Taxonomy Revised (Remembering, Understanding, Applying, Analyzing, Evaluating, Creating) to one of the ideas presented by Dr. Norvig regarding either his course, instructional practices, or student observed behaviors. Please be specific in your post when referencing the video and with which Bloom's Taxonomy Revised category you are discussing. 2) Discuss your thoughts on how what you've references in #1 above, can be applied in an online course setting.