Term | Page | Term | Page |
---|---|---|---|
Demographic Data | 65 | Content Validity | 74 |
Operational/Working Definitions | 66 | Criterion Validity | 75 |
Conceptual Definitions | 67 | Construct Validity | 75 |
Items (on an index or scale) | 71 | Reliability | 76 |
Index (scale Construction) | 71 | Split-half Reliability | 76 |
Validity | 73 | Test-Retest Reliability | 77 |
Face Validity | 74 |
Recently, I was discussing with colleagues about appropriate use of technology in a classroom. A friend of mine is starting her academic instructor career and was mentioning about how students can use/misuse technology. The main discussion hinged on the operational definition that was mentioned briefly throughout the conversation. In addition, the debate was sparked because the communication process of the operational/working definition was constantly evolving until a group consensus was reached. So, how would we define a research process? What techniques can shape our world? What techniques can change our research process?
This chapter is devoted to the proper steps and vocabulary needed to engage in a social science research process. It is meant to give you a deeper understanding of how research projects are created, evaluated as a project, and how we work to shape the project so that we can statistically evaluate what we initially set out to investigate.
In order to not skew the research, we define an item we want to study and follow the techniques in this chapter. The examples will be using the conversation from FB regarding technology use in a classroom from a sociological perspective.
Class participation: What items do we need to investigate the use/misuse of technology in a classroom from a sociological and statistical perspective? Let’s check all that apply.
Looks like something that would come from a textbook. The answers that are not related to the chapter are not correct.
Let’s take a few moments and see what we can come up with using this chapter to relate back to use/misuse of technology in a classroom. We can also change our focus and look at this in terms of a work setting. Equally applicable because we can simply change the operational definition of what we want to study. Or, we can conduct both studies and look at it in terms of academic and business settings.
Before we can even analyze the data, we have to obtain it (either by ourselves or from a repository). When you are obtaining data from a research, you must obtain IRB permissions. In this class we will be obtaining data from a repository that has a set of data for you to analyze.
background information that gives the social characteristics of a subject.
Our analysis is only as accurate as the data we obtain or we analyze.
Some instructors refer to this as a working definition. I use the term operational definition. This term is used interchangeably in the textbook.
Definition: A definition of the way that someone or something will be measured to determine the subjects score on a variable.
Considerations: Is there available data that allows us to evaluate (code) for the item(s) we want to research? Do we have enough time/money to evaluate what we want to study?
A conceptual definition is a general definition of a concept such as one would find in a textbook or dictionary. It is more specific than the operational definition that allows us to develop a research project using items to measure components of a topic we want to study.
In our case of misuse of technology in a classroom setting….
http://www.merriam-webster.com/dictionary/misuse
In our case of appropriate use of technology in a classroom setting…
In order to measure certain sociological items we use either the Likert Scale (see slides set 3) or a slight variation: the ladder question. Where on a ladder would you place yourself for enjoying a recreation event or a medical event? The type of question drives the response.
Items: various components (e.g. family values, lifestyle, etc.) to generate a scale or index.
Index: a range of scores treated as interval or ratio levels of measurement, measuring a phenomenon.
Class exercise: let’s work on page 72 for a few minutes and discuss how the type of question might generate responses.
Validity: the extent to which the concept one wishes to measure is actually being measured by a particular scale or index. In other words: are you measuring what you want to measure?
Face Validity: extent to which the measure is subjectively viewed by knowledgeable individuals as covering the concepts.
Content Validity: extent to which the measure covers all the generally accepted meanings of the concept.
Criterion Validity: the extent to which the measure is able to predict some criterion external to it.
Construct Validity: the ability of the scale to measure variables that are theoretically related to the variable that the scale purports to measure.
The likelihood that the scale is actually measuring what it is supposed to measure.
Split-half reliability: a measure of internal consistency that splits an overall scale into two scales, each containing half of the original items.
Test-retest reliability: a measure that determines an individual’s consistency in responding the same way to a specific item over time.
How likely is it that your scale is measuring what you want it to measure? We are moving into statistics at this point.
Avoid these situation that result in misleading operational definitions.
Know what you are studying!
In groups of 2 to 4, work through page 80 exercise 3.1 and 3.2. Send an email to me with the name of the students and your responses.
Note: I will be responding to the last 3 weeks of responses in a text file via email either at the end of this week or the beginning of the next week. For now, most of you are on the correct track! :)
Term | Page | Example | Understanding |
---|---|---|---|
Demographic Data | 65 | ||
Operational/Working Definitions | 66 | ||
Conceptual Definitions | 67 | ||
Items (on an index or scale) | 71 | ||
Index (scale Construction) | 71 | ||
Validity | 73 | ||
Face Validity | 74 | ||
Content Validity | 74 | ||
Criterion Validity | 75 | ||
Construct Validity | 75 | ||
Reliability | 76 | ||
Split-half Reliability | 76 | ||
Test-Retest Reliability | 77 |
Come see me during office hours or send me an email :)