Personnel selection explained

Personnel selection is the methodical process used to hire (or, less commonly, promote) individuals. Although the term can apply to all aspects of the process (recruitment, selection, hiring, onboarding, acculturation, etc.) the most common meaning focuses on the selection of workers. In this respect, selected prospects are separated from rejected applicants with the intention of choosing the person who will be the most successful and make the most valuable contributions to the organization.[1] Its effect on the group is discerned when the selected accomplish their desired impact to the group, through achievement or tenure. The procedure of selection takes after strategy to gather data around a person so as to figure out whether that individual ought to be utilized. The strategies used must be in compliance with the various laws in respect to work force selection.

Overview

The professional standards of industrial-organizational psychologists (I-O psychologists) require that any selection system be based on a job analysis to ensure that the selection criteria are job-related. The requirements for a selection system are characteristics known as KSAOs – knowledge, skills, ability, and other characteristics. US law also recognizes bona fide occupational qualifications (BFOQs), which are requirements for a job which would be considered discriminatory if not necessary – such as only employing men as wardens of maximum-security male prisons, enforcing a mandatory retirement age for airline pilots, a religious college only employing professors of its religion to teach its theology, or a modeling agency only hiring women to model women's clothing.[1]

Personnel selection systems employ evidence-based practices to determine the most qualified candidates and involve both the newly hired and those individuals who can be promoted from within the organization.[1]

In this respect, selection of personnel has "validity" if an unmistakable relationship can be shown between the system itself and the employment for which the people are ultimately being chosen for. In this way, a vital piece of selection is Job Analysis. An analysis is typically conducted before, and regularly as part of, the improvement in determination systems. Then again, a selection method may be deemed valid after it has already been executed by directing follow up job analysis and demonstrating the relationship between the selection process and the respective job.

The procedure of personnel selection includes gathering data about the potential candidates with the end goal of deciding suitability and sustainability for the employment in that particular job. This data is gathered utilizing one or more determination devices or strategies classified as such:

Development and implementation of such screening methods is sometimes done by human resources departments; larger organizations hire consultants or firms that specialize in developing personnel selection systems. I-O psychologists must evaluate evidence regarding the extent to which selection tools predict job performance, evidence that bears on the validity of selection tools. These procedures are usually validated (shown to be job relevant), using one or more of the following types of validity: content validity, construct validity, and/or criterion-related validity.

History and development

Chinese civil servant exams, established in AD 605, may be the first documented "modern" selection tests, and have influenced subsequent examination systems.[2] As a scientific and scholarly field, personnel selection owes much to psychometric theory and the art of integrating selection systems falls to human resource professionals.

In the United States of America, members of the Society for Industrial and Organizational Psychology (SIOP) conduct much of the research on selection. Primary research topics include:

Validity and reliability

Validity of interviews

The validity of interviews describes how useful interviews are in predicting job performance. In one of the most comprehensive meta-analytic summary to date by Weisner and Cronshaw (1988). The authors investigated interview validity as a function of interview format (individual vs board) and degree of structure(structure vs unstructured). Results of this study showed that structured interviews yielded much higher mean corrected validities than unstructured interviews (0.63 vs 0.20), and structured board interviews using consensus ratings had the highest corrected validity (0.64).

In[4] McDaniel, Whetzel, Schmidt & Maurer's Comprehensive Review and Meta- analysis of the Validity of Interviews (1994) paper, the authors go a step further and include an examination of the validity of three different types of interview content(situational, job-related, and psychological).Their goal was to explore the possibility that validity is a function of the type of content collected.

They define the three kinds of content as follows – situational content was described as interview questions that get information on how the interviewee would behave in specific situations presented by the interviewer. For example, a question that asks whether the interviewee would choose to report a coworker for behaving in an unethical way or just let them go. Job related questions, on the other hand, assess the interviewee's past behavior and job-related information. While psychological interviews include questions intended to assess the interviewee's personality traits such as their work ethic, dependability, honesty etc.

The authors conducted a meta-analysis of all previous studies on the validity of interviews across the three types of content mentioned above. Their results show that for job-performance criteria, situational interviews yield higher mean validity(0.50) than do job-related interviews(0.39) which yield a higher mean validity than do psychology interviews(0.29). This means that when the interview is used to predict job performance, it is best to conduct situational interviews rather than job-related or psychological interviews. On the other hand, when interviews are used to predict an applicant's training performance, the mean validity of job-related interviews(0.36) is somewhat lower than the mean validity of psychological interviews(0.40).

Going beyond the content of the interview, the authors' analysis of interview validity was extended to include an assessment of how the interview was conducted. Here, two questions emerged – Are structured interviews more valid than unstructured interviews ? and are board interviews(with more than one interviewer) more valid than individual interviews.

Their answer to the first question – Are structured interviews more valid unstructured interviews was that structured interviews, regardless of content, is more valid(0.44) than unstructured interviews(0.33) in predicting job performance criteria. However, when training performance is the criteria, the validity of structured and unstructured interviews are similar (0.34 and 0.36).

As for the validity of board interviews versus individual interviews, the researchers conducted another meta-analyses comparing the validity of board interviews and individual interviews for job performance criteria. The results show that individual interviews are more valid than board interviews(0.43 vs 0.32). This is true regardless of whether the individual interview is structured or unstructured.

When exploring the variance in interview validity between job performance, training performance, and tenure criteria, the researchers found that the interviews are similar in predictive accuracy for job-performance and training performance(0.37 vs 0.36). But less predictive for tenure (0.20).

Validity of cognitive ability and personality tests

Based on meta-analysis results, cognitive ability tests appear to be among the most valid of all psychological tests and are valid for most occupations. However, these tests tend to do better at predicting training criteria than long term job performance. Cognitive ability tests in general provide the benefit of being generalizable. Hence they can be used across organizations and jobs and have been shown to produce large economic gains for companies that use them (Gatewood & Feild, 1998; Heneman et al., 2000).

But despite the high validity of cognitive testing, it is less frequently used as selection tools. One main reason is that cognitive ability testing has been demonstrated to produce adverse impact. In general, groups including Hispanics and African-Americans score lower than the general population while other groups including Asian – Americans score higher (Heneman et al., 2000; Lubenski, 1995). The legal issues with cognitive ability testing were amplified by the supreme court's ruling in the famous 1971 Griggs v. Duke Power case. In this case, the Supreme Court ruled that when a selection test produces adverse impact against protected group members the company must be able to defend it by showing that use of the test is a "business necessity" for the operation of the business. The courts have held narrow interpretations of business necessity that require companies to show that no other acceptable selection alternative exists (Sovereign, 1999). As a result, many companies abandoned cognitive ability testing (Steven L. Thomas &Wesley A. Scroggins, 2006).

While the utility of cognitive ability testing in selection has been broadly accepted, the utility of personality testing, until relatively recently, has not. Historically, research documenting the low predictive validity and the potential for invasion of privacy based on item content has made its application as selection instruments questionable (Hogan, Hogan, & Roberts, 1996).

But due to the legal challenges associated with cognitive ability, interest in personality instruments has recently been revived (Schmidt, Ones, & Hunter, 1992). Some have suggested that pairing personality testing with cognitive ability testing may be one means to enhance validity while reducing adverse impact (Ryan, Ployhart, & Friedel, 1998). Because it is very likely that some aspects of personality enhance individual ability to apply intellectual capacity while other personality traits limit its application (Kaufman & Lichtenberger, 1999). Hence, adding a personality to an ability test should enhance validity while reducing the adverse impact of a selection system.

Recent research studies prove this assumption to be false, by showing that the addition of a predictor producing smaller group differences (i.e., personality test) to a predictor producing higher group differences (i.e., cognitive ability test) does not reduce the potential for adverse impact to the degree that is often expected (Bobko, Roth, & Potosky, 1999; Schmitt, Rogers, Chan, Sheppard, & Jennings, 1997).

Although the use of personality tests with measures of cognitive ability may not have the desired effects on reducing adverse impact, it appears that the addition of personality measures to measures of cognitive ability as a composite predictor results in significant incremental validity (Bobko et al., 1999; Schmitt et al.,1997). These studies found that the validity of predictor composites was highest when alternative predictors were used in combination with cognitive ability. Though this combination of predictors resulted in the highest predictive validity, the inclusion of cognitive ability with these alternative predictors increased the potential for adverse impact (Steven L. Thomas & Wesley A. Scroggins, 2006).

In summary, cognitive ability testing by itself has been shown to have high levels of validity, but comes with issues relating to adverse impact. Personality tests, on the other hand, have historically been proven to have low validity due to the lack of a common understanding on what constitutes personality, and the non-standardized measures available. But a growing number of research studies show that the best way for organizations to achieve close to optimal validity and job performance prediction, is to create a predictor composite that includes a measure of cognitive ability and an additional measure such as a personality test.

Predictor validity and selection ratio

Two major factors determine the quality of newly hired employees, predictor validity (aka predictive validity) and selection ratio.[1] The predictor cutoff is a test score differentiating those passing a selection measure from those who did not. People above this score are hired or are further considered while those below it are not. If the test accurately differentiates between the successful and unsuccessful workers using this test score cutoff then it is high on predictor validity.

The selection ratio (SR), on the other hand is the number of job openings n divided by the number of job applicants N. This value will range between 0 and 1, reflecting the selectivity of the organization's hiring practices. When the SR is equal to 1 or greater, the use of any selection device has little meaning, but this is not often the case as there are usually more applicants than job openings. Finally, the base rate is defined by the percentage of employees thought to be performing their jobs satisfactorily following measurement.

Selection decisions

Tests designed to determine an individual's aptitude for a particular position, company or industry may be referred to as personnel assessment tools. Such tests can aid those charged with hiring personnel in both selecting individuals for hire and in placing new hires in the appropriate positions. They vary in the measurements they use and level of standardization they employ, though all are subject to error.[5]

Predictors for selection always have less than perfect validity and scatter plots, as well as other forecasting methods such as judgmental bootstrapping, and index models can help us to refine a prediction model as well as identify any mistakes.[1] [6] The criterion cutoff is the point separating successful and unsuccessful performers according to a standard set by the hiring organization. True positives are applied those thought to succeed on the job as a result of having passed the selection test and who have, in fact, performed satisfactorily. True negatives describe those who were correctly rejected based on the measure because they would not be successful employees.

False negatives occur when people are rejected as a result of selection test failure, but would have performed well on the job anyway.[1] Finally, false positives are applied to individuals who are selected for having passed the selection measure, but do not make successful employees. These selection errors can be minimized by increasing the validity of the predictor test.

Standards for determination of the cutoff score vary widely, but should be set to be consistent with the expectations of the relevant job.[1] Adjusting the cutoff in either direction will automatically increase the error in the other. Thus, it is important to determine which type of error is more harmful on a case-by-case basis.

Banding is another method for setting cutoff values.[1] [7] Some differences in test scores are ignored as applicants whose scores fall with in the same band (or, range) are selected not on the basis of individual scores, but of another factor spas to reduce adverse impact. The width of the band itself is a function of test reliability, the two being negatively correlated. Banding allows employers to ignore test scores altogether by using random selection, and many have criticized the technique for this reason.

Predicting job performance

A meta-analysis of selection methods in personnel psychology found that general mental ability was the best overall predictor of job performance and training performance.[8]

Regarding interview procedures, there are data which put into question these tools for selecting employees.[9] While the aim of a job interview is ostensibly to choose a candidate who will perform well in the job role, other methods of selection provide greater predictive validity and often entail lower costs. Unstructured interviews are commonly used, but structured interviews tend to yield better outcomes and are considered a better practice.[10]

Interview structure is defined as "the reduction in procedural variance across applicants, which can translate into the degree of discretion that an interviewer is allowed in conducting the interview."[11] Structure in an interview can be compared to a typical paper and pencil test: we would not think it was fair if every test taker were given different questions and a different number of questions on an exam, or if their answers were each graded differently. Yet this is exactly what occurs in an unstructured interview; thus, a structured interview attempts to standardize this popular selection tool.

Multiple studies and meta-analyses have also been conducted to look at the relationship between organizational citizenship behavior (OCB) and organizational performance and success.[1] [12] Job candidates exhibiting higher levels of helping, voice, and loyalty behaviors were generally rated as more confident, received higher salaries, and received higher salary recommendations than job candidates exhibiting these behaviors to a lesser degree.[12] This was found to be true even candidate responses regarding task performance were taken into account. Finally, content analyses of open-ended question responses indicated selection decisions were highly sensitive to candidates with low expression of voice and helping behaviors.

See also

External links

Notes and References

  1. Muchinsky, P. (2012). Psychology Applied to Work, (10th ed.). Summerfield, NkluC: Hypergraphic Press.
  2. Têng. Ssu-yü. Têng Ssu-yü. Chinese Influence on The Western Examination System. 2717830. Harvard Journal of Asiatic Studies. 1943. 7. 4. 267–312. 0073-0548. 10.2307/2717830.
  3. Steel, P. (2006). From the work one knows the worker: A systematic review of the challenges, solutions, and steps to creating synthetic validity. International Journal of Selection and Assessment, 14(1), 16–36.
  4. Comprehensive Review and Meta- analysis of the Validity of Interviews. McDaniel, Whetzel, Schmidt & Maurer's. 1994. Journal of Applied Psychology . 79 . 4 . 599–616. 10.1037/0021-9010.79.4.599. 12324059 .
  5. U.S. Department of Labor Employment and Training Administration. (1999) "Chapter 1: Personnel Assessment." pp. 1–7 http://www.saylor.org/site/wp-content/uploads/2011/07/psych304-3.1.pdf
  6. Armstrong. J. Scott. others. 2012. Predicting job performance: The moneyball factor. Foresight: The International Journal of Applied Forecasting. 25. 31–34.
  7. Campion, M. A., Outtz, J. L., Zedeck, S., Schmidt, F. L., Kehoe, J. F., Murphy, K. R., & Guion, R. M. (2001). The controversy over score banding in personnel selection: Answers to 10 key questions. Personnel Psychology, 54(1), 149–185.
  8. Schmidt, F. L. & Hunter, J. E. (1998). The validity and utility of selection methods in personnel psychology: Practical and theoretical implications of 85 years of research findings. Psychological Bulletin, 124(2), 262–274.
  9. McDaniel, M. A., Whetzel, D. L., Schmidt, F. L., & Maurer, S. D. (1994).The validity of employment interview: A comprehensive review and meta-analysis . Journal of Applied Psychology, 79(4), 599–616.
  10. Huffcut, A. I. (2010). From science to practice: Seven principles for conducting employment interviews . Applied H.R.M. Research, 12, 121–136.
  11. Huffcut, A. I., & Hunter, W. Jr. (1994). Hunter & Hunter revisited: Interview validity for entry-level jobs. Journal of Applied Psychology, 79, 184–190.
  12. Podsakoff, N. P., Whiting, S. W., Podsakoff, P. M., & Mishra, P. (2011). Effects of organizational citizenship behaviors on selection decisions in employment interviews. Journal of Applied Psychology, 96(2), 310–326.