Sunday, 10 November 2013

KILLING OF HAKIMULLAH MEHSUD: ITS IMPLICATIONS ON PAK US REALTIONSHIP

ALTAF AHMAD DAR US Pakistan-relationship continuing over 65 years, both countries have sustained it, sometimes happily, other times not so happily. Many see the U.S. strategic interests are inextricably linked with a stable Pakistan. President Dwight D. Eisenhower famously called Pakistan America’s “most allied ally in Asia.” Relations between the United States and Pakistan have been seriously strained several times over recent years, including in 2011, when U.S. forces killed al Qaeda mastermind Osama bin Laden in a raid that Pakistan said violated its sovereignty. But cash-strapped Pakistan depends to a great extent on U.S. support. Washington, despite frustrations over the relationship, is unlikely to break completely with its nuclear-armed ally. The Pakistani Prime Minister’s October 2013 state visit to Washington was viewed by many as a good response to a recent deterioration in U.S.-Pakistan ties.” The recently improved relationship got once again twist on 1 Nov.2013 by the death of Tehrik-e-Taliban (TTP) Pakistan leader Hakimullah Mehsud in a US drone that could cripple the group but undermine an effort by Pakistan’s government to engage militants in peace talks.It has infuriated the Pakistani government and Interior Minister Chaudhry Nisar Ali Khan said that Pakistan will review relations with the US and make complaints over the American attack to the UN Security Council. "This is not just the killing of one person. It's the death of all peace efforts." Imran Khan, the cricketer-turned-politico who heads the Tehrik-i-Insaf party (PTI), has also condemned the timing of the strike, saying the U.S. has been thwarting the peace process in Pakistan with the ongoing CIA secret drone campaign in the country. In retaliation, he has proposed to block NATO supplies to Afghanistan that run through the province where his party is in power. Killing of Mehsud will impact on many dimensions of Pak US relationship. The first is that Pakistan is the main route for supplies for U.S. troops in landlocked Afghanistan, for everything from food and drinking water to fuel, and the closure of the routes could be a serious disruption as U.S. and other Western forces prepare to withdraw from Afghanistan by the end of next year. Pakistani cooperation is also seen as vital in trying to bring peace to Afghanistan, in particular in nudging the Afghan Taliban, allied but separate from the Pakistani Taliban, into talks with the Kabul government. Second, Mehsud is considered as one of the brave and respectful leader among TTP and it could affect the peace talks between govt and TTP. One of The TTP Taliban leader criticized the government and alleged that on one hand it is talking about peace talks with the Taliban and on the other it has failed to stop the US drone strikes. He said talks could not be held with the government when Taliban leaders were being targeted in American drone strikes. It would be now tough for Pakistan to start peace talks with the Talban and it is considered that the leaders who are in the race of Mehsud’s Succession are tough and there would be little hope of compromise. Even if talks started, it was unclear how successful they would be unless the government gave significant concessions to the militants. Third the public outcry against the drone strikes in the Pakistan will intensify, which Pakistan is believed to be an attack on its sovereignty. In Last month, Mr Sharif in a visit to Washington repeated his request for an end to the drone strikes. Pakistan has learned few lessons from its past intimacy with the US. Islamabad’s number one complaint vis¬ ଠvis the U.S. has always been, and is, that Washington has proven to be an untrustworthy partner and all this shows the extreme pace with which intolerance is growing in Pakistani society. What is needed at this point of time is that the US should stop drone attacks in Pakistan, take into full confidence on the war on terror and all day to day progress in Afghanistan. ALTAF AHMAD DAR IS A RESEARCH SCHOLAR SCHOOL OF EDUCATION PONDICHERRY UNIVERSITY daraltaf87@gmail.com

Saturday, 26 October 2013

Teacher effectiveness, concept, principles and criteria



Before knowing the meaning of teacher effectiveness, it is essential to know the meaning of effectiveness. Effectiveness is a term used extensively by different people in different contexts. So it is defined in different ways. Teacher education and job performance of a teacher are the contexts in which this term is used. Effectiveness are the requirements of a competency based teacher education, which includes knowledge, skills and values the trainee teacher, must demonstrate for successful completion of the teacher education programme
Anderson (1991) stated that “… an effective teacher is one who quite consistently achieves goals which either directly or indirectly focus on the learning of their students”.
Dunkin (1997) considered that teacher effectiveness is a matter of the degree to which a teacher achieves the desired effects upon students. He defined teacher competence as the extent to which the teacher possesses the requisite knowledge and skills, and teacher performance as the way a teacher behaves in the process of teaching. The term “teacher effectiveness” is used broadly, to mean the collection of characteristics, competencies, and behaviours of teachers at all educational levels that enable students to reach desired outcomes, which may include the attainment of specific learning objectives as well as broader goals such as being able to solve problems, think critically, work collaboratively, and become effective citizens. A few characteristics of a competency are as follows. 
1. A competency consists of one or more skills whose mastery would influence the attainment of the competency.
2. A competency has its linkage with all the three domains under which Performance can be assessed. It covers the domains of knowledge, skill and attitude.
3. Competencies, since have a performance dimension of them, are observable and demonstrable.
4. Because the competencies are observable, they are also measurable. It is possible to assess a competency from the performance of a teacher. It is not necessary that all competencies of a teacher have the same extent of knowledge, skill and attitude. There may be some competencies of a teacher which have the same extent of knowledge, skill and attitude. There may- be some competencies involving more of knowledge than skill and attitude, whereas, same competencies may be skill/performance loaded.


Who is an effective teacher?

Effective teachers are those who achieve the goals which they set for themselves or which they have set for them by others (Ministries of Education, Legislators and School administrators). As a consequence, those who study and attempt to improve teacher effectiveness must be cognizant of the goals imposed on teachers or the goals that teachers establish for themselves or both. A corollary of this definition is that effective teachers must possess the knowledge and skills needed to attain the goals, and must be able to use that knowledge and those skills appropriately if these goals are to be achieved. The possession of knowledge and skills falls under the heading of Teacher competence. In contrast, the use of knowledge and skills in the classroom is referred to as teacher performance. Thus those who investigate and attempt to understand the teacher effectiveness must be able to link teacher competence and teacher performance with the accomplishment of teacher goals (that is teacher effectiveness). Four major assumptions are implicit in this definition of teacher effectiveness. The first is that effective teachers tend to be aware of and actively pursue goals. These goals, in turn, guide their planning as well as their behaviours and interactions with students in the classroom. The second assumption is that teaching is an intentional and reasoned act. Teaching is intentional because we always teach for some purpose, primarily to facilitate learning. Teaching is reasoned because what teachers teach their students is judged by them to be worthwhile. A third assumption implicit in this definition of teacher effectiveness is that the vast majority of teachers goals are, or should be, concerned either directly or indirectly with their students learning. An example of direct teacher concern with learning is a teacher who states that he or she intends to help students develop the ability to differentiate facts from opinions, or reality from fantasy. An example of indirect teacher concern with learning is a teacher who sets out to decrease the level of disruptive behaviour in the classroom because he or she believes that learning cannot occur before the level of disruptive behaviour is reduced. A fourth assumption underlying this definition of teacher effectiveness is that no teacher is effective in every aspect of their profession. For example a primary school teacher may be highly successful in teaching reading comprehension to his or her students while struggling to teach the elements of rudimentary problem solving in mathematics.



Characteristics associated with more effective Teachers
Professionalism:
Commitment: Commitment to doing everything possible for each student possible and enabling all students to be successful.
Confidence: Belief in one’s ability to be effective and to take on challenges
Trustworthiness: Being consistent and fair, keeping ones word.
Respect: Belief that all individuals matter and deserve respect.
Thinking/Reasoning
Analytical thinking: Ability to think logically, break things down and recognise cause and effect.
Conceptual thinking: Ability to identify patterns and connections, even when a great deal of detail is present.
Expectations:
Drive for improvement: Relentless energy for setting and meeting challenging targets for students and the school.
Information seeking: Drive to find out more and get to the heart of things, intellectual curiosity.
Initiative: Drive to act now to anticipate now and pre-empt events.
Leadership:
Flexibility: Ability and willingness to adapt to the needs of a situation and change tactics.
Accountability: Drive and ability to set clear expectations and parameters and hold others accountable for performance.
Passion for learning: Drive and ability to support students in their learning, and help them become confident and independent learners.  
Criteria for effective teaching
Product variable: It means what students learnt. Persons who prefer this criterion think that the best test of teacher effectiveness is how much and how well students achieve. This includes achievement in all domains-cognitive, affective and psychomotor. Examples of this kind of criterion are.
  1. Students gain in a subject matter, knowledge and related abilities.
  2. Students gain in psychomotor skills and related abilities.
  3. Students gain in interests, attitudes, personality integration, and other effective characteristics. It can be directly assessed with tests and performance ratings before, during and after instruction.
Process variable: In this variable judgment about teacher effectiveness are made in terms of what the teacher does, what the students do, the interactions between student and teacher, or all the three, but not in terms of student achievement. Examples of this kind of criterion are:
  1. Teacher behaviours, such as explaining, questioning, leading a discussion, counselling evaluating etc.
  2. Student behaviours, such as courtesy, industriousness, attentiveness, conducting an experiment, leading a discussion.
  3.  Student-teacher interactions, such as teacher-directed and student-directed exchanges, information exchanges and warmth. The use of process criterion involves making observations of teacher and student behaviours during instruction in the classroom.
Presage variable: Presage is partly a predictive factor. A teachers present and probable future effectiveness is judged on the basis of his intellectual ability, college grades, personal appearance and other characteristics. Examples for this kind of criterion are intellectual abilities of the teacher, amount of college work completed in the teaching, major grade point average in college, personal characteristics and others. This can be assessed indirectly from college records, tests and ratings outside the classroom.   
Sources for deriving teacher competency
Effective teachers consistently achieve goals that focus on desired outcomes for their students. Teacher effectiveness is encompassed in knowledge, attitudes, and performance.
1. Knowledge
Teachers have excellent verbal and written communication skills.Teachers have thorough knowledge of the subjects they teach and pedagogical methods for teaching those subjects to students. Teachers know a variety of pedagogical strategies, and when and with which students these are appropriate and likely to be effective. Teachers have a thorough understanding of the linguistic and cultural backgrounds of their students and how best to maximize learning for students with diverse needs and characteristics. .Teachers knows how to organize and manage classrooms, using time effectively. .Teachers know how to assess student learning, both formally and informally, and how to vary instruction for students based on these assessments. Teachers know how to select and make resources that are appropriate for student learning activities.Teachers understand language development and children’s developmental stages at the level they teach.
2. Attitudes
Teachers respect their students regardless of their background, language, or ethnicity Teachers use formative evaluation to adjust instruction and diversify it for the needs of individual or groups of students. Teachers create warm and caring relationships with their students. Teachers maintain collaborative relationships with their peers and with parents and community members. Teachers have high expectations for the learning of all students. .Teachers view student errors as a window to their thinking that can be used to improve student learning. Teachers are reflective about their practice. Teachers believe in collaboration with others toward common goals for student learning. .Teachers are receptive to involvement of parents and community members in their classrooms. Teachers are eager to continue to learn and to improve their practice. Teachers are committed to their profession.
3. Performance
Teachers’ classrooms are well organized, providing an environment that fosters an interest in learning. Teachers develop classroom rules with students and maintain safe and orderly classrooms in which all students are treated fairly and equitably.Teachers make effective use of time, both of overall classroom time and the time of individual students. Teachers use effective teaching techniques: planning lessons, presenting new material clearly, helping students connect new learning with previous learning, and providing guided and independent practice for new material taught. .Teachers provides opportunities for students to be actively involved in their own learning. .Teachers respond to student errors in positive ways that help students understand and learn the concepts involved. Teachers use formative evaluation to adjust instruction and diversify it for the needs of individual or groups of students.Teachers create warm and caring relationships with their students. Teachers maintain collaborative relationships with their peers and with parents and community members.
4. Planning
Teachers' planning refers to that aspect of teaching where teachers formulate a course of action. It is an activity that is typically carried out in the absence of students and before the actual teaching. Teacher’s plans, which serve as 'scripts' (whether they are done on paper or in one's mind), include decisions on what to teach ad how to teach the chosen content.


5. Motivation
Even when the plan is good, an important function of the teacher in implementing it, is to motivate those learners who are demotivated o the task of learning and nurture those who are already well motivated. There are several ways in which one can achieve this.
6. Presentation and communication
After ensuring the students' interest in the learning, a teacher in the classroom is to transact with the students in the context of a specific subject matter. The teacher is expected to communicate with the students in a number of ways so that the learners attain various types of learning outcomes. In order to achieve this effectively, the teacher may have to manifest various types of skills including lecturing, explaining, eliciting through questions, conducting discussion, dramatizing, reading, demonstrating; using audio-visual aids, etc. all these may be categorised into skills for effective presentation and communication in the instructional situation.
7. Evaluation
Evaluation of the students' achievement of a pre-specified objective is part and parcel of a teacher's function, what would these evaluation skills include? Preparing question papers? Taking viva voce? Yes, all these are included in the evaluation process. But these are only a part of the total evaluation function of a teacher. He has to observe the students in many different situations in order to judge the extent to which the expected terminal behaviours have been actually achieved by them. This includes so many activities. A teacher has to; first of all, select the suitable techniques and tools of evaluation. It is obvious that one cannot measure the length of a stick by using a weighting machine. Similarly, for measuring the skills of performing experiments, one cannot have a written examination. Hence the teacher has to select the suitable techniques and tools for measurement according to what he would like to measure.
Problems for measuring the teacher effectiveness

1. The basic format of the traditional “teacher effectiveness” study is a so-called product-process design, similar to that used in most school effectiveness research. In particular, it does not take into account sufficiently the fact that teachers' roles are broader than their classroom practice and includes management roles, pastoral roles, and relationships with parents and community as well as classroom practice.
2. A major problem in classroom observation is observer bias. If an observer has a preconceived expectation of the behaviour of the observed person, s/he is likely to align the actual observations to these expectations to at least some extent. In one example, students received information that a visiting lecturer was either a warm or a cold personality. Following the lecture, they were asked to rate the lecturer. The half of the class who had received the information that he had a warm personality rated him a more effective, sociable, humorous, and informal lecturer than did the half who had received the information that he was a cold person. Careful training can help alleviate this problem.
3. Generalisability has been found to be a problem in observational research, in that behaviours observed on one occasion are not necessarily typical of a teacher's everyday behaviour. This points to the need to observe teachers as many times as possible and the need to take contextual factors into account.
4. Some individuals have argued that student achievement is the most logical criterion on which to evaluate teachers. We reject this superficially attractive notion for three reasons: student variability, the regression effect, and the limitations of currently available achievement tests.
 5. Rating scales lack the minimum properties necessary or accurately measuring the performance of teachers. Second, such rating scales lack validity. And third, they are highly susceptible to the halo effect. We question the validity of rating scales and of teacher e valuations based on them, because such scales reflect the beliefs of the raters about the nature of competent performance -not actual teaching competence.


Reflections:
It was a very good experience for me to work on teacher effectiveness and competency. From the beginning I faced problems to understand the topic but thanks to madam who gave me moral support to work on this topic. I learned meaning of teacher effectiveness and the requirements of a competency based teacher education.




EDUCATIONAL INTEGRATION IN SOUTH ASIA


EDUCATIONAL INTEGRATION IN SOUTH ASIA South Asian Association for Regional Cooperation (SAARC) countries have common features, such as geographic and climatic conditions, and share issues concerning the socioeconomic, cultural, and educational advancement of their people. Education affects every area of national development and deserves requisite attention. All SAARC countries have a similar education structure, including entry qualifications and age, duration of courses, and instructional management system. But all the south Asian countries are facing problems like access, low enrolment, low enrolment of girls, low GNP on education and substandard quality of education. It is the need of the hour to south Asian countries to evolve common educational forums and cooperative strategies to deal with the situation. Mutual cooperation is a key factor in dealing with these problems. SAARC aims to accelerate economic and social development in member states, which requires optimal utilization of human resources. It has, therefore, emphasized the importance of promoting education. At SAARC's ninth summit at Male in 1997, it was acknowledged that illiteracy was a major factor impeding development of human resources of South Asia and contributed significantly to the region’s socioeconomic backwardness. Cooperation in education entered the SAARC agenda early with the establishment of a Technical Committee on Education in 1989. At the Thirteenth SAARC Summit held in Dhaka in November 2005, the leaders noted the achievements of the Member States during recent years in the area of primary education and stressed that to meet the challenges of the twenty-first century Member States must make important strides in the areas of science, technology and higher education. SAARC has signed a MOU with UNESCO in 2008 with the aim to achieve education for all. At the invitation of the Government of Sri Lanka, the First Meeting of the SAARC Ministers of Education/Higher Education was held in Colombo on 27 March 2009 preceded by the Meeting of the Senior Officials of the Ministries of Education/Higher Education. At the Thirteenth SAARC Summit held in November 2005 in Dhaka, India proposed to create a Centre of Excellence, in the form of a South Asian University, which can provide world class facilities and professional faculty to students and researchers drawn from every country of the region. And ultimately SAARC University was established in 2011 in Delhi. South Asian countries have signed many agreements from time to time and some of the important among them are the SAARC Chair, Fellowship and Scholarship Scheme instituted in 1987, SAARC Consortium on Open and Distance Learning in 1999, committee of Heads of University Grants Commission/Equivalent Bodies in 1999 and the SAARC Agenda for Culture was finalised on 9 Dec. 2006. Nevertheless, a lot more remains to work before concrete benefits of such cooperative activities are clearly visible


EDUCATIONAL INTEGRATION IN SOUTH ASIA

South Asian Association for Regional Cooperation (SAARC) countries have common features, such as geographic and climatic conditions, and share issues concerning the socioeconomic, cultural, and educational advancement of their people.  Education affects every area of national development and deserves requisite attention.  All SAARC countries have a similar education structure, including entry qualifications and age, duration of courses, and instructional management system. But all the south Asian countries are facing problems like access, low enrolment, low enrolment of girls, low GNP on education and substandard quality of education. It is the need of the hour to south Asian countries to evolve common educational forums and cooperative strategies to deal with the situation. Mutual cooperation is a key factor in dealing with these problems. SAARC aims to accelerate economic and social development in member states, which requires optimal utilization of human resources. It has, therefore, emphasized the importance of promoting education. At SAARC's ninth summit at Male in 1997, it was acknowledged that illiteracy was a major factor impeding development of human resources of South Asia and contributed significantly to the region’s socioeconomic backwardness.

Cooperation in education entered the SAARC agenda early with the establishment of a Technical Committee on Education in 1989. At the Thirteenth SAARC Summit held in Dhaka in November 2005, the leaders noted the achievements of the Member States during recent years in the area of primary education and stressed that to meet the challenges of the twenty-first century Member States must make important strides in the areas of science, technology and higher education. SAARC has signed a MOU with UNESCO in 2008 with the aim to achieve education for all. At the invitation of the Government of Sri Lanka, the First Meeting of the SAARC Ministers of Education/Higher Education was held in Colombo on 27 March 2009 preceded by the Meeting of the Senior Officials of the Ministries of Education/Higher Education.  At the Thirteenth SAARC Summit held in November 2005 in Dhaka, India proposed to create a Centre of Excellence, in the form of a South Asian University, which can provide world class facilities and professional faculty to students and researchers drawn from every country of the region. And ultimately SAARC University was established in 2011 in Delhi. South Asian countries have signed many agreements from time to time and some of the important among them are the SAARC Chair, Fellowship and Scholarship Scheme instituted in 1987, SAARC Consortium on Open and Distance Learning in 1999, committee of Heads of University Grants Commission/Equivalent Bodies in 1999 and the SAARC Agenda for Culture was finalised on 9 Dec. 2006. Nevertheless, a lot more remains to work before concrete benefits of such cooperative activities are clearly visible


RESEARCH SCHOLAR SCHOOL OF EDUCATION PONDICHERRY UNIVERSITY daraltaf87@gmail.com

Thursday, 20 September 2012

Research Methodology


26 July 2011 AA3RESEARCH METHODOLOGY SEMESTER-I Teaching Notes Unit-1: Aim and Scope of Research in Education Research-meaning, importance and characteristics 1. Define Educational Research? Research is an inquiry process that has clearly defined parameters and has as its aim the discovery of creation of knowledge, or theory building; testing, confirmation, revision, refutation of knowledge and theory; and /or investigation of a problem for local decision- making. C.r.McClure and P.Hersson, 1991,p.3 Scientific research is a systematic, controlled, empirical and critical investigation of hypothetical propositions about the presumed relations among natural phenomena. F,N.Kerlinger, 1978, p.11 Research is a planned, systematic and caution endeavor undertaken to find out something new or to deepen or to better our understanding about something that we have already known. The ‘new’ in our finding may be a new phenomenon (and knowledge about it) or uses or applications of a phenomenon (or knowledge) or solutions to problems that we face. A better (more correct) understanding will take us nearer to reality and lead to better, wider and more forms of applicability of the phenomenon(knowledge)being researched. Research is a continuous endeavor. Its findings are based on probabilities. They are acceptable rationally within the boundaries of the School of philosophy and empirically with reference to the evidences collected. When rationality becomes more comprehensive and evidences support a higher level of facts, research advances further towards reality. Thus research is a continuous progressive process. It is performed to identify, understand and apply the naturally occurring physical or socially constructed phenomena. It generates theory, promotes productive applications, and improves practices; it may lead to discoveries and/or applications. ______________________________________________________________________________ Dr.K.Chellamani, 26 July 2011 2. Write the importance of educational research.(Need) Research in education is essential for the development of education as a science and for improving the system of education. Precisely, it is needed for (i) investigating the nature of the children, (ii) investigating the learning process and motivation, (iii) developing curricula, (iv) developing better learning materials and procedures,(v) refining teaching procedures, (vii) improving management, (viii) predicting '‘academic success’, (ix) investigating the causes of undesirable phenomena in education,(x) evaluating effectiveness of different educational ‘treatments’, and so on. Education has strong roots in the fields like philosophy, psychology, sociology, economies and history. It is through an intensive process of scientific enquiry about the impact of these disciplines on various aspects of education that sound theories of education can be established. Sound education is recognized as basic towards the development of individual, social, and national development. It helps in improving policies, systems and practices of the government. Educational practitioners are always in search of effective methods of instruction, more reliable techniques of evaluation, richer learning materials, more efficient systems of management, etc. Decisions based on systematic research in education save time, money, and energy, and reduce failure, and show us the path of progress. It is because they are scientifically tested, and their worth well-established. Research in education is assuming greater importance and urgency due to the rapid expansion and democratization of education throughout the world. It becomes the right of every individual in India. It results in sensitization of innumerable problems like problems of expansion, inclusion, buildings, finance, strategies, individual differences, media and materials, managementetc. Extensive quality research is needed to solve all these problems. Problems also arise in the phenomenon like academic achievement, discipline, wastage and stagnation, truancy, mass copying in examinations, etc. This calls for investigations and remedial measures. In the world of Technology, there comes a device every day. A lot of research is needed for its effective usage, development of efficiency among people, and productivity. And above all, the concept of education changes every day. According to it, the education system has to be modified, for which research is needed. ______________________________________________________________________________ Dr.K.Chellamani, 26 July 2011 3. List the main characteristics of educational research: On examining the definitions, the characteristic of educational research goes as- Research is an inquiry or investigation; It has clearly defined parameters; It describes the problems; It explains the problems; It generalizes the findings; It predicts the phenomena; It is a critical inquiry; It is a systematic inquiry; It is time-bound; It should satisfy the external and internal validity; It is objective and transferable; It is a controlled inquiry; and It is an empirical inquiry. 4. What are the AIMS of educational research? • Finding truth: Research is a search for truth. • Testing Hypothesis(es): Whenever a research worker is confronted with a probable relation known as hypothesis(es), he/she then, through scientific process, tests the hypothesis(es). • Establishing relationship(s): Many a time researches are undertaken to study the relationship between some variables, for example relationship between the socio-economic status of parents and the achievement of their off-springs. • Arriving at their generalization or theory building: On the basis of research findings, generalizations are arrived at or theories are built, like theories of learning. • Verifications and testing of existing facts, theories, generalizations, etc.(Improvement of knowledge): Research findings also help in verifying the existing theories or generalizations arrived at by other researchers. • Revision of existing theories: The existing theories undergo revision; whenever contracting evidences come to light. • Finding solutions to problems/questions: Research is also a problem solving activity in a systematic way. • Providing a scientific basis for planning, decision making, policy formulation, policy revision, etc.: The database generated through research activity is helpful in the decision-making and planning process. • Predicting behaviors, forecasting outcomes/events on the basis of research, viz., population projections, number of teachers and schools needed in the next decade, etc. ________________________________________________________________________ Dr.K.Chellamani, 29 July 2011 Teaching Notes 5. Give an account of the Areas of Educational research. Research is a highly purposeful and systematically conducted activity. Every piece of research has its own clearly defined objectives. It is to answer certain questions or solve certain problems. It leads to confirmation or discovery of new relationships, principles, other theoretical propositions and they may have their implications and applications. Thus research generates theory, promotes productive applications, and improves practices; it may lead to discoveries and/or applications. Research in Educationis essential for the development of education as a science and for improving the process/system of education. Research is much needed for (1) investigating the nature of children, (2) investigating the learning process and motivation, (3) developing curricula,(4)developing better learning materials and procedures, (5) refining teaching procedures,(6) refining evaluation procedures, (7) improving management, (8) predicting academic success,(9) investing the causes of undesirable phenomena in education, (10) evaluating effectiveness of different educational ‘treatments’, and so on. There are certain other considerations that emphasize the need for and importance of research in education and they are as follows.  Education has strong roots in the fields like philosophy, psychology, sociology, economics and history. There is an impact of these disciplines on education. An intensive scientific enquiry on its impact may establish sound theories of Education.  Sound education is basic for individual, social and national development. Therefore research is required to improve educational policies, systems and practices. Educational practitioners are constantly in search of effective methods of instruction, more reliable techniques of evaluation, richer learning materials, more efficient systems of management, etc. There is dignity on the part of educational practitioners in using research proved practices.  The rapid expansion and democratization of education throughout the world necessitates research in education. Since the right of every individual to fulfill development through education has been recognized everywhere, every country is aiming at providing universal education to its citizens in the shortest possible period of time. Consequently, a number of educational problems- problems of expansion, buildings, finance, strategies, individual differences, media and materials, management, etc have arisen.. For successful solutions, extensive research of quality is needed.  Many undesirable phenemenoa like poor academic achievement, indiscipline, wastage and stagnation, truance, mass copying in examinations, etc, are observed in the field of education. Investigations of these problems with a view to devising remedial measures is badly needed.  The expansion and innovation in technology calls for research for its appropriate usage and improve education. It is required to improve quality of education, and also to economise effort and increase efficiency.  There is also need for research because of the changing conception of education reflecting in the commission reports. “Indeed, in the 21st century everyone will be required to demonstrate independence, judgment and more personal responsibility if common objectives are to be reached. Our report also underlines another requirement, namely that none of the talents lying dormant like hidden treasures in every individual should be allowed to go unused. These talents, to name but a few, include: memory, logical thought, imagination, physical ability, an aesthetic sense, the ability to communicate and the natural charisma of a group leader. In actual fact these abilities only serve in underlining the importance of more self-knowledge (...).The "information revolution" is increasing ease of access to information and facts. This means that education should be striving to make it possible for everyone to collect and single out relevant information, to put it in context, to handle information and to use it. To this end, then, education should seek to adjust itself constantly to the changes going on in society; it must be able to pass on the acquisitions, the foundations and the richness of human experience." Dr.K.Chellamani, 30 August 2011 Teaching Notes M.Ed. Teaching Notes What is Qualitative and Quantitative Research? Since ages, man tend to have grips with his environment and to understand the nature of the phenomena it presents to his senses. The means by which he sets out to achieve these ends may be classified into three broad categories: experience, reasoning and research( Mouley, 1978). Unlike the other two, scientists construct their theories carefully and systematically out of their research. The hypotheses they formulate are being tested empirically so that their explanations have a firm basis. They are explanations based on the data collected. Beyond the veracity of the data, there is a question of its significance. Here comes qualitative research. It answers the 'what' and 'why' of the problem. Creswell's(1998) list of qualitative research is named as “five traditions”. Those traditions are, in alphabetical order: biography, case study, ethnography, grounded theory, and phenomenology. The list can be added with other researches such as evaluation, action research and critical theoretical analysis. Shank (2002) distinguishes qualitative research with quantitative research with a metaphor. Quantitative research was likened to the metaphor of the window. Windows are used to give us a clear and transparent look at things. When researchers are trying to be clear and transparent, such issues as reliability, validity, bias, and “generalizability” rise to the surface. Windows are most useful in giving us clear, typical, broad, accurate and realistic views of things in the empirical world. Qualitative research, on the other hand, was likened to the metaphor of the lantern. Lanterns are used to allow light to illuminate dark areas so that we can see things that previously were obscure. Once we shed light on things, we understand them better. Considering the above explanations, we take up the following as the essential criteria for a researcher to abide by. (1) depth in investigation, (2) adequate interpretation, (3) fertile illumination, and (4) accountability in participation. Investigation is the duty of the researchers. They must be aware of the fact that any phenomena in the empirical world has much remained below the current awareness. The investigator has to go below the surface to discern matters and issues that only reveal themselves under careful scrutiny i.e. the depth he has to go in. That is, the researchers are not to be contented with familiar perspectives and preconceptions. Once they are convinced that they have gone into the areas previously unaddressed with enough depth to the notice of the readers, then comes the issue of adequate interpretation. The investigator has the responsibility of forming more complete and complex understanding of those things with an intelligible and manageable grasp of the phenomena under examination. Once the researchers have investigated with depth and created interpretative understanding for newer insights, the next question is ' how to use it?'. Hence the researcher must illuminate the new ways of looking and the new ways of practicing it. Researchers participate in a variety of procedures, and they must be accountable for any and all forms of participation. Researchers must document these efforts in order to see how they have interacted and what some of the anticipated and unanticipated fruits of such anticipation might be. The logic of empirical research must be consistent for qualitative and quantitative research. Quantitative research differs from the qualitative in its clear framework. Whereas qualitative research has a mixed method. The hypotheses testing in the quantitative research may be more powerful when comparing to the description and narration of the qualitative research. That does not mean that the mixed-method studies are not important and that should not be done. The important fact is they are actually a part of quantitative research to complement the hypothesis testing. The flexibility in the qualitative research is replaced by an array of systematic techniques in quantitative research. It has rigorous design of its own. There is identification and control of explicit variables. Whereas the qualitative research do not allow too much preplanning and this may diminish the power and potential usefulness of the study. There is a belief that all elements in a qualitative research should be coded. But if the researcher is not careful, too much data become a threat for minute details. Coding is for laying out key issues. They have the informed freedom to sort through their data, looking for the important details that help in explaining the phenomena in a newer way. They are more concerned with tactical and flexible exploration rather than with creating reductive and scientific coding and thematic structures. They may be able to address these issues as they arose in practice. The above discussion leads us to follow the conditions listed below. (1) Careful examination of research assumptions. The researcher must be aware of the implicit assumptions. The qualitative researchers are to make the strange familiar and even prepare to make the familiar strange. They must move into those dark and unexamined areas. This cannot be possible by adhering to old and established ways of doing things. (2) Researchers must be ready to change their directions and must convey later why and how the change was made. (3) Education is a lively process. Hence the research process - the research design and the presentation of report should be lively. Then that research will live forever. Validity and Reliability Validity: It is an important factor in a research. If a piece of research is invalid then it is worthless. So it is the worthiness of a research. It is the requirement for both qualitative and quantitative research. It is defined as it measures what it purports to measure. In case of qualitative research, it is addressed through the honesty, depth, richness and scope of the data achieved, the participants approached, the extent of triangulation and the objectivity of the researcher. In quantitative research, data validity may be improved through careful sampling, appropriate instrumentation and appropriate statistical treatment of the data. It is impossible for hundred percentage validity. Quantitative research possesses a measure of standard error which is inbuilt and which has to be acknowledged. In qualitative research there is a degree of bias in the subjectivity of the respondents, their opinions, attitudes and perspectives. Therefore validity should be seen as a matter of degree rather than as an absolute state (Gronlund, 1981). Hence the researcher should strive to minimize invalidity and maximize validity. There are several kinds of validity and here we see the four important validities. They are, content validity, construct validity, concurrent validity and face validity. In fact validity is the touchstone of all types of educational research, shown in generalizability, replicaplity and controllability. Maxwell, echoing Mishler (1990) suggests that 'understanding' is more a suitable term than 'validity' in qualitative research. It means validity attaches to accounts, not to data or methods (Hammersley and Atkinson, 1983); it is the meaning that subjects give to data and inferences drawn from the data that are important. 'Fidelity”( Blummenfeld-Jones, 1995) requires the researcher to be as honest as possible to the self-reporting of the researched. Data selected must be representative of the sample, the whole data set, the field i.e., they must address content, construct and concurrent validity. Researchers go for instruments to obtain information for solving the research problem in hand. If a teacher wants to know about the impact of the new curriculum on students’ achievement, he needs both an instrument to record the data and some sort of assurance that the information obtained will enable him to draw correct conclusions. The drawing of correct conclusions based on the data obtained by use of an instrument is what validity is about. Validity can be defined as referring to the appropriateness, meaningfulness and usefulness of the specific inferences researchers make based on the data they collect. The researcher has to validate an instrument before he takes it for testing. It is the degree to which evidence supports any inferences a researcher makes based on the data he collects using a particular instrument. It is not validating the instruments but rather the inferences about the specific uses of an instrument itself. These inferences should be appropriate, meaningful, and useful. Appropriateness refer to relevance i.e, to the purposes of the study. If the purpose of the study is to know about students understanding on the spelling rules, it would make no sense to make inferences about this from their scores on a test about the grammar in English. Meaningfulness says about the meaning of the information( such as test scores) collected through the use of an instrument. The researcher has to see what exactly a high score on a particular test mean; what does that score tell us about the person who receives that score; and in what way this person differs from a person who receives a low score. These information are collected from people. These are warranted conclusions about the people on whom the data were collected. Usefulness of the instrument helps researchers to make decisions related to what they were trying to to find out. Researchers interested in knowing the dropout rates, for example, need information that will enable them to infer the cause behind it. Validity therefore depends on the amount and type of evidence the researcher has on which he makes interpretations. The following are the four types of validity a researcher has to establish before administrating a tool. Content validity: This is the first key a researcher has to attend when he develops an instrument for his study. The instrument must cover the domain or items that it purports to measure. If the domain has a broad width, a fair representation of the wider issue can be taken. Care should be given to sampling of items with representativeness. That is the elements of the main issue to be covered in that particular research are both a fair representation of the wider research and the elements chosen are addressed in depth and breath. For example, if a researcher wanted to find out a section of students' ability in spelling the words in English around 1000. Due to certain factors, he decided to take only 50 words for testing. Then that test should ensure representativeness of the 1000 words, may be in terms of spelling rules. It refers to the content and format of the instrument. How appropriate is the content? How comprehensive it is? How adequate are the items to represent the content to be assessed? The content and the format must relate to the definition of the variable and the sample of the subjects to be measured. Construct validity: A construct is an abstract. It is on the operationalization of the tool. Every item should mean what the researcher proposed to assess. That is the articulation of the construct. In the above example, the test should focus only on the spelling rules and nothing other than this. If the tool is on testing intelligence, it should be on that particular intelligence what the researcher expected the student to demonstrate. To establish construct validity, the researcher must see, – whether his construction of a particular issue agree with other constructions of the same particular issue( it could be done through correlations with other measures of the issue) – the theory of what that construct is and its constituent element is – also counter examples which might falsify his construction. This exercise of confirming and refuting evidence allow the researcher to stand clearly and acknowledge conflicts if any. In qualitative research, construct validity must demonstrate that the categories that the researchers are using are meaningful to the participants themselves (Eisenhart and Howe, 1992:648)i.e., that they reflect the way in which the participants actually experience and construe the situations in the research; that they see the situations through the actors' eyes. Campbell and Fiske(1959) and Brock-Utne (1996) suggest that convergent validity implies that the different methods for researching the same construct should give a relatively high inter correlation, whilst discriminant validity suggests that using similar methods for measuring different constructs should yield relatively low inter correlations. Construct validity is to find out - How well does this construct explain differences in the behavior of individuals or the performance on certain tasks. Hence the articulation of the construct is very important. If the researcher wants to assess the aptitude of students, he should look into the acceptable, expected ability on that particular component at that particular age. He should know the theory of what that construct is and its constituent element. Then going for evidences to confirm and refute help him in establishing construct validity. Criterion-related validity: It is to relate the results of one particular instrument to another external criterion. It has two principal forms: Predictive validity and Concurrent validity. Predictive validity is the validity of any particular data which proved to be the same in the future data of the same sample. For example, the entrance test examination results have its reflection in their performance in their professional course performance. Concurrent validity goes with the meaning of concurrence. That is, the data gathered from one instrument must correlates highly with data gathered from using another instrument. Both are similar in its core concept i.e, agreement with a second measure. They differ in one point, where the absence of time element in the former. Concurrence can be demonstrated simultaneously with another instrument. It is to see how strong is the relationship and how well do such scores estimate or predict future performance of a certain type? Ensuring validity: It is very important to maintain validity at every stage of research. The researcher has to take up confidence in building validity in the elements of the research plan, data acquisition, data processing analysis, interpretation and its ensuing judgment. Louis Cohen, Lawrence Manion and Keith Morrison(2000) have given a detailed account of minimizing threats to validity at every stage of investigation. It is given below. At the design stage: ✗ choosing an appropriate time scale; ✗ ensuring that there are adequate resources for the required research to be undertaken; ✗ selecting an appropriate methodology for answering the research questions; ✗ selecting appropriate instrumentation for gathering the type of data required; ✗ using an appropriate sample(e.g. One which is representative, not too small or too large); ✗ demonstrating internal, external, content, concurrent and construct validity; 'operation alizing', the constructs fairly; ✗ ensuring reliability in terms of stability(consistency, equivalence, split-half analysis of test material); ✗ selecting appropriate foci to answer the research questions; ✗ devising and using appropriate instruments (for example to catch accurate, representative, relevance and comprehensive data(King, Morris and Fitz-Gibbon, 1987); ensuring that readability levels are appropriate; avoiding any ambiguity of instructions, terms and questions; using instruments that will catch the complexity of issues; avoiding leading questions; ensuring that the level of test is appropriate- neither too easy nor too difficult; avoiding test items with little discriminability; avoiding making the instruments too short or too long; avoiding too many or too few items of each issue; ✗ avoiding a bias choice of the researcher or research items(insiders or outsiders as researchers). At the stage of data gathering: ✗ reducing the Hawthorne effect; ✗ minimizing reactivity effects (respondents behaving differently when subjected to scrutiny or being placed in new situations, for example in the interview situations- we distort people's lives in the way we go about studying them (Lave and Kvale, 1995:226)); ✗ trying to avoid dropout rates among respondents; ✗ taking steps to avoid non-return of questionnaires; ✗ avoiding having too long or too short an interval between pre-tests and post-tests; ✗ ensuring inter-rater reliability; ✗ matching experimental and control groups fairly; ✗ ensuring standardized procedures for gathering data or for administrating tests; ✗ building on the motivations of the respondents ✗ tailoring the instruments to the concentration span of the respondents and adressing other situational factors(e.g. Health, environment, noise, distration and threat); ✗ addressing factors concerning the researcher(particularly in a interview situation); for example the attitude, gender, race, age, personality, dress, comments, replies, questioning technique, behavior, style and non-verbal communication of the researcher. At the stage of Data analysis: ✗ Using respondent validation; ✗ avoiding subjective interpretation of data(e.g. Being too generous or too ungenerous in the award of marks),i.e. Lack of standardization and moderation of results; ✗ reducing the halo effect, where the researcher's knowledge of the person or knowledge of other data about the person or situation exerts an influence on subsequent judgments; ✗ using appropriate statistical treatments for the level of data (avoiding applying techniques from interval scaling to ordinal data or using incorrect statistics for the size, type and complexity, sensitivity of data); ✗ recognizing spurious correlations and extraneous factors which may be affecting the data(i.e. Tunnel vision); ✗ avoiding poor coding of qualitative data; ✗ avoiding making inferences and generalizations beyond the capability of datato support such statements; ✗ avoiding the equating of correlations and causes; ✗ avoiding selective use of data; ✗ avoiding unfair aggregation of data(particularly of frequency tables); ✗ avoiding unfair telescoping of data(degrading the data); ✗ avoiding Type I and/or Type II errors. At the stage of data reporting: ✗ avoiding using data very selectively and unrepresentativly (for example accentuating the positive and neglecting or ignoring the negative); ✗ indicating the context and parameters of the research in the data collection and treatment, the degree of confidence which can be placed in the results, the degree of context-freedom or context-boundaries of the data)i.e. The level to which the results can be generalized); ✗ presenting the data without mis-representing their message; ✗ making claims which are sustainable by the data; ✗ avoiding inaccurate or wrong reporting of data (i.e. technical errors or orthographic errors); ✗ ensuring that the research questions are answered; releasing the results neither too soon nor too late. Having the understanding over the areas of invalidity, the researcher can take steps to ensure the minimization of errors to the maximum. Reliability It refers to the consistency of the scores obtained. That is, how consistent they are for each individual from one administration of an instrument to another, from one set of items to another. For example in a test to measure the word power of the students- If the test is reliable, we would expect a student who receives a high score the first time he takes the test to receive a high score the next time he takes the test. The scores obtained from an instrument can be quite reliable, but not valid. Suppose a researcher gave two forms of a test on Psychology to a group of B.Ed students and found their scores to be consistent. Those who scored high on Part A also scored high on Part B; and those who scored low on Part A scored low on Part B; and so on. Here the scores are reliable. Then, if the researcher used the same test scores to predict the success of the students in Language classes, he would be surprised to see that there is no relation. Any inferences about success in Language based on scores on psychology would have no validity. It is understood that the scores which are inconsistent for a person are not valid and hence they do not provide any useful information. Errors of Measurement: In the case of giving the same test twice to the sample, the scores or answers may not be identical. Probably, it is due to a variety of factors (differences in motivation, energy, anxiety, a different testing situation, and so on), but it is inevitable. Such factors result in errors of measurement. Since errors of measurement are always present to some degree, researchers expect some variation in test scores ( in answers or ratings, for example) when an instrument is administered to the same group more than once, when two different forms of an instrument are used, or even from one part of an instrument to another.. Reliability estimates provide researchers with an idea of how much variation to expect. Such estimates are usually expressed as another application of the correlation coefficient known as a reliability coefficient. A validity coefficient expresses the relationship that exists between scores of the same individuals on two different instruments. A reliability coefficient also expresses a relationship, but this time it is between scores of the same individuals on the same instrument at two different times, or between two parts of the same instrument. The three best known ways to obtain a reliability coefficient are the test-retest method; the equivalent-forms method; and the internal consistency methods. Unlike other uses of the correlation coefficient, reliability coefficients must range from 0.00 to 1.00. .(Fraenkel,1993). Test-Retest Method It is the administration of the same test twice to the same group after a certain time interval has elapsed. In order to calculate the reliability coefficient, the relationship between the two sets of scores is obtained. Here we have to look into the time factor. The longer the time interval, the lower the reliability coefficient as there is a greater likelihood of changes in the individuals taking the test. To check the evidence of test=retest validity, an appropriate time interval should be selected. This interval should be that during which individuals would be assumed to retain their relative position in a meaningful group. It is not possible to study a variable that has no stability. It is natural that some abilities (such as writing) are more subject to change than others (such as abstract reasoning). At the same time some personal characteristics such as self-esteem are considered to be more stable than others (such as teen age interests). Scientists say, for educational research, stability of scores over a two to three month period is usually viewed as sufficient evidence of test-retest reliability coefficients, therefore, the time interval between the two testing should always be reported. Equivalent- forms Method It is the administration of two different but equivalent or alternative or parallel forms of an instrument to the same group of individuals. The questions are not the same but they sample the same content and they are constructed separately from each other. A reliability coefficient is then calculated between the 2 sets of scores obtained. A high coefficient would indicate strong evidence of reliability- that the two forms are measuring the same thing. It is possible to combine the test-retest and equivalent forms methods by giving two different forms of the same test with a time interval between the two administrations. A high reliability coefficient would indicate not only the two forms are measuring the same sort of performance but also what we might expect with regard to consistency over time. Internal-consistency methods The above methods involve administration of two tests or testing two sessions. There are methods which show internal consistency, estimating reliability, by administrating a single instrument. The two important methods are Split-Half procedure and Kuder-Richardson approach. Split-Half procedure It involves scoring of items into two different halves, viz. odd and even. Every individual’s score will be taken in these halves and then applied correlation coefficient for two set of scores. The coefficient indicates the degree to which the two halves of the test provide the same results, and hence describes the internal consistency of the test. The Spearman-Brown prophecy formula is used to calculate the reliability coefficient. The formula goes as follows: Reliability of Scores on total test = 2 x reliability for ½ test ________________________ 1 + reliability for ½ test Hence a correlation of coefficient of 0.56 is obtained by comparing one half of the test items to the other half, the reliability of scores for the total test would be: Reliability of Scores on total test = 2 x 0.56 1.12 _________ = ________=0.72 1 + 0.56 1.56 The above illustration implies an important characteristic of reliability. The reliability of a test (or any instrument) can generally be increased by increasing its length if the items added are similar to the original ones. Kuder-Richardson approach It is the most frequently employed method for determining internal consistency. It has two formulae – KR20 and KR21. If the items are of equal difficulty KR21 can be used.( The formula KR20 does not require the assumption that all items are equal difficulty.)It requires only three pieces of information, the number of items in the test, the mean, and the standard deviation. KR21 reliability coefficient = K _______ [1- __M(K-M)_____ K-1 K(SD) Here K refers to the number of items in the test, M stands for the set of test scores, and SD stands for standard deviation of the set of test scores. Alpha Coefficient Alpha Coefficient is frequently called as Cronbach alpha (after the man who developed it). This coefficient is a general form of the KR20 formula to be used in calculating the reliability of items that are not scored right versus wrong, as in some essay tests where more than one answer is possible. Scoring Pattern Generally, instruments are administered with specific directions and are scored objectively. There may not be any room for his/her judgment, though there may be differences between scorers to certain extent. This is not the case with essay evaluations. Instruments that use direct observations are vulnerable to observer differences. Such instruments go with the degree of scoring agreement. INSTRUMENTATION Researches are based on collection of data. The findings and conclusions are based on what the data reveal. Hence researchers should take are on the kinds of data to be collected, the methods of collection to be used and the scoring of the data. Let us see what data means. Data refers to the the kinds of information researchers obtain on the subjects,viz.- ✔ Demographic particulars ✔ scores of test conducted ✔ responses in the oral interview ✔ written responses to the a survey questionnaires ✔ essays written by students ✔ performance reports ✔ anecdotal records and so on. The whole process of collecting data is called instrumentation. It involves selection of items, design of the instruments and the conditions under which the instrument will be administered. Moreover the location of the data collection, the time of collection, the frequency of administration and the person to administer. The above fats are important so as to make the data reliable and useful. Thereby the researcher can draw accurate conclusions about the characteristics of the people being studied. Validity, Reliability,Objectivity and Usability A validity of an instrument means that it measures what it purports to measure. It is in the hands of the researchers how they present their inferences from the data collected through the instrument they have used. Therefore they go for the instrument which help them to draw valid conclusions about the characteristics of the subjects they study. For example to measure the reading ability of the subjects the instrument should measure such achievement. Reliability is the other important factor to be taken care of. It is on the consistency of the results. If that researcher tested the reading abilityof that particular subjects at two or more different times, he should get almost the same result every time. Then that tool has reliability. Objectivity refers to the absence of subjective judgments. The researcher has to avoid subjectivity, when gives judgments on the performance or the characteristic of the subjects. Usability is how easy it will be to use that particular instrument which he has developed; how long will it take to administer ; are the directions indicated clearly; is it appropriate for the group to whom it will be administered; how easy the scoring pattern is; how easy it is to interpret; does it have equivalent forms; does it have evidence of reliability and validity and the like. If a researcher attends all the above questions carefully, he can do justice to the research at the end. Means of Data collection instruments Once the instrument is ready, now comes the question of obtaining information. It could be the researcher himself without getting any assistance, or directly from the subjects or from other informants. 1) Researcher instruments: Eg. Rating scales, Interview schedules, Tally sheets, Flow charts, Performance check lists, Anecdotal records and Time and motion logs 2) Subject completes: Eg. Questionnaires, Self-check lists, attitude scales, Personality or character inventories, achievement/ Aptitude tests, Performance tests, Projective devices and socio-metric devices 3) Informer instruments: Eg.Rating scale, anecdotal records and Interview schedule ( Classification of instruments can be 1. Written-response-type instruments, 2. Performance-type-instrument) Source of the instrument: Either by using the previously existing instrument of some sort or by developing on his own. Unobtrusive measures: It refers to data collection procedures that involve no intrusion into the naturally occurring course f events. It may be in the form of record keeping. Types of scores: Raw scores and Derived scores Norm-referenced and Criterion-referenced Instruments Measurement scales( Nominal, ordinal,Interval scales, Ratio scales) ____________ Research Methods Historical Research: The Purpose: 1. To aware past failures and success 2. Learn how things were done/ application 3. To assist in prediction (Policy Makers). 4. To test hypo concerning relate or trends. Eg. Secondary school teachers have enjoyed greater prestige than elementary school teachers since 1940. 5. To understand present educational practice and policies more fully. Steps: 1. Defining the problem or question to be investigated. 2. Locality relevant sources of historical information. 3. Summarizing and evaluating the information obtained from theses sources. 4. Presenting and interpreting the information as it realated to the problem or question the originated study. Defining the problem: Describe, clarify and explain and sometimes to correct (as when a researcher finds previous accounts of an action or event to be in error). Eg. Better to study a well – defined problem in depth i.e. Perhaps more narrow than one told like than to pursue a more broadly stated problem then cannot be sharply defined or fully resolved. Titles: Eg: • The decline in age at leaving home, 170 – 2000. • Annadurai and his speeches. • Kamaraj and his political principles. • Women professionals before independence. Locating relevant Resources: Categories of Sources: Documents, numerical records, oral statements and records and relics, songs, stories and legends. A relic is any object whose physical or visual character is can provide some information about the past. Eg. Furniture, art work, clothing, buildings, monuments or equipment. Primary Vs Secondary Sources: Primary Source: Who was a participant in or a direct witness to the event. Secondary Source: A document prepared by an individual who was not a direct witness to an event but who obtained the description of the event from someone else. Eg: A newspaper editorial commentary on a Teacher’s speech. Summarizing information Edward J. Carr (Historian) Evaluation of his Key questions for any historical researches • Was this document really written by the supposed author? (i.e. is it genuine?) [External Criticism] • Is the information contained in the document true? (i.e. is it accurate?) [Internal Criticism] External Criticism: (Nature and authenticity of the document itself) • Who wrote • For which purpose • When was it written? • Where was it written? • Under what condition • Do different forms or versions of the document exist? Internal Criticism: (What the document says) These are the data presented (Attendance records, Budjet features, test scores and so on) reasonable? With regarding to the author of the document: • Was the author present at the event or she is describing? • Was the author a participant in or an observer of the event? • Was the author competent to describe the event (refer to the qualification of the author) • Was the author emotionally involved in the event? • Did the author have any vested interest in the outcomes of the event? With regarding to the contents of the document: • Do the contents make sense? • Could the event described have occurred at that time? • Would people have behaved as described? (A major danger is – presentism – describing present day beliefs, values and ideas to people who lived at another time – a historical hindsight) • Does the language of the document suggest a bias of any sort? (Emotionally charged, intemperate of the author) • Do other versions of the event exist? Generalization in Historical researches: - Rarely able to study an entire population of individuals or events. - A sample of the phenomena of interest. Disadvantages: 1. No control over threats. 2. Limitations are severe (document analysis). 3. Cannot ensure representationer’s analysis. 4. nor can they check validity and reliability of the inferences made from the data available. 5. The possibilities of researcher’s bias exist. 6. Any observed relationships are due to subject, character, implementation, history, maturation, attention or location. 7. Depends mostly on the skill and integrity of the researcher – since methodological controls are unavailable – it is the different research method. Case Study Analysis • Case study analysis can penetrate situations in ways that are not always susceptible to numerical analysis. • It establishes cause and effect. The strength in that they observe the effects in real contexts, recognizing that contact in a profile determinant of both causes and effects. • Case study approach is valuable when the researcher has little control over events. Hallmarks: • Concerned as a rich and vivid description of events relevant to the case. • Provides a chronological narrative of events to the case. • Blends a description of events and analysis of them. • Focuses on individual actors or group of actors and seeks to understand their perception of events. • It highlights specific events that are relevant to the case. • The researcher is integrally involved in the case. • An attempt is made to portray the richness of the case in writing up the reports. • It can make theoretical statements but like other researches must be supported by the evidence. Types: (Yin, 1984) a. Exploratory (as a pilot to other studies or research of rations) eg. Discovery Channel b. Descriptive (Providing narrative accounts) c. Explanatory (Testing theories) a can be used to generate hypotheses that are tested in large scale surveys, experiments or other forms of research. Eg. Observational. The important of study to be faced in undertaking care studies. According to Adelmen et al, 1980; Nisbet and Watt, 1984; Hitchcock and Hughes, 1995, 1. What exactly is a case? 2. How are cases identified and selected? 3. What kind of case study is this? (Purpose) 4. What is reliable evidence? 5. What is objective evidence? 6. What is an appropriate selection to include from the wealth of generated dates? 7. What is a fair and accurate account? 8. Under what circumstances is it fair to take an exceptional case? 9. What kind of sampling is most appropriate? 10. To what extent is triangulation required and how will this be addressed? 11. What is the nature of validation? 12. How will the balance be struck between uniqueness and generalization? 13. What is the most appropriate form of writing up and reporting case studies? 14. What ethical issues are exposed in undertaking a case study? A key issue in Case study research is the selection of information. - Need not always adhere to representativeness ________, unrepresentative but critical incidents or events occur that are crucial to the understanding of the case. Significance rather than frequency is a hallmark of Case Studies offering the researcher an insight into the real dynamics of situations and people. Two ways/ Types of Case Studies: Participatory and Non-participatory observation. Non-participatory observation: Eg. Flander’s I Analysis. Advantages in Participatory observation: 1. Observations are superior when data are being collected on non-verbal. 2. Able to make appropriate notes about its salient features. 3. Bee of its extended period of time researchers can develop more intimation and informal relationships. 4. Case Study observations are less reactive unless others in answering questionnaires. Designing Case Studies Yin(1994) identified 5 components. 1. A Study’s question. 2. Its proposition, if any. 3. Its unit(s) of analysis. 4. The logic linking the data to the propositions. 5. The criteria for interpreting the findings. - How and why of study – their definition is the first task of the research. - An exploratory study would have a purpose. The unit of analysis defines whether the case is linking the data to propositions and the criteria for interpreting the findings are the vast developed aspects in the case studies. Campbell (1975) described “Pattern – matching”, it is for internal validity – only in explanatory case. External validity is in generalizable beyond the immediate case. Reliability is achieved in many warp case studies. One such is development of the case study protocol. • An overview of the case study project (Objectives, issues, topics being investigated). • Field procedures (Credentials and access to sites, sources of information) • Case study questions (specific questions – investigator must keep in mind during data collection) • A guide for case study report (Outline, format for the narrative) 6 sources of evidence: - Documents (Performance data) - Archival records (Service records, original list of names etc.,) - Interviews - Direct observation - Participant observation - Physical artifacts Mills and Hukeman (1954) suggested analytic techniques such as - Rearranging the arrays - Placing the evidence in the matrix of categories - Creating flow charts or data displays - Tabulating the frequencies of different events - Using mean variances and cross tabulations to examine the relationship between variables and other such techniques to facilitate analysis. 4 parts in analysis 1. Show that the analysis relied on all the relevant evidence. 2. Include all major rival interpretation in the analysis. 3. Address the most significant aspect of the cast. 4. Use the researcher’s prior, expert know to further analysis. • Representationer: in comparison, one in generalizing to a theory based on cases selected to represent dimensions of that theory. • Case selection should be theory driven. • Cross – theoretic case selection. • Pattern matching – to establish – to find out quantitative and consistency qualitative evidence. • Process tracing - a. Controlled observation (On key variables) Time – series analysis (Not just before and after observation but outside the range of normal fluctuation of time series) • Congruence testing (Suitableness)/ (What are identical agreement cases) when there are a large number of cases it could ne statistical methods of correlation and control. • Explanation – building. Triangulation. Ethnographic Research or Naturalistic inquiry: • Ethnic means pertaining to the custom, dress, food etc., of a particular racial group or cult. • Ethnography means the scientific descriptions of the races of the earth. • Ethnographic means pertaining to ethnography. • Combination of participant and non participant observation. • Obtain a holistic picture of a particular society, group, institution, settling or situation. • Portraying the everyday experiences of observing and interviewing. • Seldom initial to precise hypothesis. Eg. What is life like in an international public school? Making use of all the tools to get data. Advantages: 1. A comprehensive perspective. 2. Research topics those are not easily quantified. 3. Study a group behavior over a time. Disadvantages: 1. No way to check the validity of the researcher’s conclusions. As a research observers bias is impossible to eliminate. 2. Since – a single situation – no generalizability. 3. No hypo – so relationships between variables is unclear. 4. Pre planning and review are less useful. So pitfalls in the methodology are unidentified. Risk prone. 5. Time consuming. Ethnography is suitable for the following: • Those by nature defy simple quantification. • Those best be understood in a natural setting. • Those involve the study of individual or group activities over time. • Roles of the educators play and behaves associated to the roles. • Study of activities of a key. • Involving the study of formal organization. Ex: 1. Rural – teacher 2. Questioning at home and at school; a comparative study. Research must include ‘thick’ descriptions. Researchers are the instruments of the research. Attribution of meaning is continuous overtime. Researchers generate rather than test hypotheses. Theory generation is derivative – grounded – the data suggested the theory rather than vice versa. • Purposive sampling enables the full scope of issues to be explored. • Research design emerges over time. • Applications are tentative and pragmatic. • Trustworthiness and its components replace more conventional views of reliability and validity. • Needs propositional knowledge and tacit knowledge. • The researcher becomes the human instrument in the field building on her tacit knowledge in addition to her propositional knowledge. Using methods that fit comfortably to human inquiry. Eg. Observation, interviews, documentary analysis, unobtrusive methods. The advantage of her research is her adaptability, responsiveness, knowledge, ability to handle sensitive matters, ability to set the whole picture, able to clarify the summarize to explore, to analyze, to examine typical or idiosyncratic responses. • Ethnography has extensive use of unstructured observations, conversations, documented by detailed field notes from basis for this type of research. • It attempts to explain the uses of interdependence of group behaviors and interactions. Strategies to enhance Trustworthiness: 1. Bracketing in the process of the researcher becoming self – aware and reflecting on the research process and her own assumption. 2. Prolonged engagement and persistent observation. 3. Multiple data sources. It will provide opportunities for comparison of data among and bit participants as well as different types of data sources. 4. Participant checking. (follow up and feed back) 5. Peer researcher support group throughout the research process, support froup members review and comment on transcripts of participant observation, interviews, discuss memos written by the researcher and provides a forum for discussing researcher’s ideas. 6. Conformability and dependability of the results – rich description. Stage 1. Locality a field of study. 2. Addressing ethical issues. 3. Deciding the sampling. 4. Finding a role and managing entry into the context. 5. Finding information. 6. Developing and maintaining relations in the field. 7. Data collection in situation. 8. Data collection outside the field. 9. Data Analysis. 10. Learning the field. 11. Writing the report. Observation: • Like data from ‘live’ situations. • To understand the context of the program to be open – ended and inductive, to see things that might otherwise be unconsciously missed. • Discover things. • Access personal knowledge. • Observation enables the researcher to gather data analysis. • The physical setting. • The human setting. • The interactional setting. • The program setting. From unstructured to structured. From responsible to pre-ordinate (Worked out in advance). A semi – structured – in a far less systematic manner. Hypothesis generating From complete participation to a detachment Closeness and distance Familiarity and strangeness Quantitative research tends to have a small field of focus, fragmenting the observed into minute chunks that can subsequently be aggregated into a variable. Qualitative research – Situation unfold ‘Fitness of Purpose’ Structured: the research will need to decide 1. The foci of the observation (eg. People as well as the event) 2. The frequency of the observations (eg. Every 30 seconds, every minute, every 2 minutes) 3. The length of the observations period (eg. 1 hour, 20 minutes) 4. The nature of the entry (1 coding system) Student to student Student to students; Student to teacher; Students to Teacher; Teacher to Students; Teacher to Student; Student to self; Task in hand; Previous task; Future task; Non – task. (Every 30 seconds) There are 4 principal ways of entering data onto a structured observation schedule; Event Sampling: Teacher shouts at the child I I I I I Child shouts at the child I I I Parents shouts at the teacher I I Teacher shouts at the parents I I Instantaneous sampling: otherwise called time sampling: If it is important to know the chronology of events, then it is necessary to use Instantaneous sampling, observation at standard intervals of time. Interval Recording Entry at fixed intervals. However instead of charting what is happening on the instant, it charts what has happened during the preceding interval. It enables frequencies to be calculated simple pattern to be observed and an appropriate sequence of events to be noted, because it charts what has taken place in the preceding interval of time. Rating Scale: 5 point scale 1 – not at all 2 – very little 3 – a little 4 – a lot 5 – a very great deal Why Observation: Delude – deceive, cause to accept what is false or true. 1. We may delude ourselves about what is happening. We observe what we want to study, harmonious relationship; effective practice, rules that are consistently followed. Content Analysis Analysis the content of speeches, essays, documents, journals, films, music and others. It is a research technique for the objective, systematic and quantitative description of the manifest content of communication. Procedure: 1. Specification of objectives. Indicates the purpose. Eg. School text book – it it is affective objectives in the lessons the categories of such objectives listed and categorized accordingly. 2. Hypotheses formulation derived after the review of the research. Research hypotheses should not be written in the inner form and should predict single difference or relationship. Each term in hypothesis should be clearly defined and should be testable. 3. Sampling specific samples of content. There are 3 universal from that which samples may be drawn (i) titles (a specific Npap, a lesson in a text book or journal), (ii) issues, (iii) content found in issues or titles. Eg. Sports in newspaper. 4. Determining the categories; According to Bateson, (i) ‘What is said’, (ii) ‘How it is said’ (i) includes ‘subject matter’ (i.e., what the communication is about) Direction, standards, values, methods, traits, actor, authority, origin and target. (all of them, refer to the ‘wh or whi’ categories. Ie., information or facts) (ii) ‘form’ or type of communication, form of statement, intensity and device included denoting higher order thinking. 5. Category Analysis: As per the nature and purpose of the investigation the categories are analyzed into components of sub categories. 6. Quantification: determining the frequency of what each of the units. 7. Standardizing the procedure: the coding procedures must be tried the persons who actually do the categories. Validity reliability and objectives need to be ensured when coding is done. 8. Reliability: The accuracy of coding determines the reliability of quantitative data. In some instances, inter observer reliability is established, by the extent to which a number of coders can consistently analyzed the same data. Another method of establishing such reliability is by having one or more observers coding the same data over a period of time. Serious limitation concerning the reliability of qualitative material should be avoided, when coding is done. Otherwise, the investigatory findings will suffer from several defects. Content analysis thus offers a research technology for document analysis. 1. Definition and categorization of Unit: The universe of the content is to be analyzed. It is a reflection of the theory or hypotheses being tested. It spells out, in effect, variables of the hypothesis. 2. Units of Analysis: Berelson lists 5 major units of analysis; words; themes; characters; items and space and time measures. a. Eg. Value words and non – value words. Difficult medium and lazy works. b. The theory is useful but more difficult unit. Eg. Discipline – an interesting larger themes, child training. c. Characters and space – and – time measures. Character – number of inches, pages, paras, number of matters of discussions and so on. d. Items – is important an essay, news story, tele – program, class recitation or discussion, autobiography. Quantification: 1. Nominal measurement – frequency. 2. ranking – ordinal measurement judges may asked to rank according to specific criteria. 3. rating – children’s pieces can be rated for degrees of creativity, originality, achievement orientation interests, values and other variables. Before quantification check whether they are representative and to count carefully, the category items are in sufficient numbers. Otherwise, generalization from statistics calculated from them is unwarranted. Selection of materials for analysis is of theory do not fulfill. The conduction those can be used for suggestive purposes and not for relating variables to each other. Survey Method: Cross – sectional longitudinal Steps: 1. Problem definition 2. Identification of the target people 3. Mode of data collection i. Direct administration ii. Mail survey iii. Telephone survey iv. Personal interview 4. Selection of the instrumentation 5. Preparation of the instrumentation 1. is this a question that can be asked exactly the way it is written. 2. Is this a question that will mean the same thing to everyone? 3. Is this a question that people can answer. 4. Is this a question that people will be willing to answer, given the data collection procedures? 5. Types of questions: Enclosed – ended of - Unambiguous - Keep the focus as simple as possible - Of short - Use common language - Avoid the use of the words that might ‘bias’ responses. - Avoid leading questions - Avoid double questions 6. presenting the questionnaire Preparation of the cover letter. Threats to the validity of instrumentation process. Location, history, instrumentation.