1 / 14

Improving the Sailor-Rating Match by Considering Interests: JOIN

2. Operational Problem. Navy classification process fails to encourage enlistment and combat attritionCurrently emphasizes short-term recruiting quotas over Sailor-rating matchDoes not consider interest/job satisfaction as key variablesMakes 20-30 year career decisions based on 7-10 minute interview with Navy Classifier.

calix
Download Presentation

Improving the Sailor-Rating Match by Considering Interests: JOIN

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


    1. Improving the Sailor-Rating Match by Considering Interests: JOIN William L. Farmer, Ph.D.

    2. 2 Operational Problem Navy classification process fails to encourage enlistment and combat attrition Currently emphasizes short-term recruiting quotas over Sailor-rating match Does not consider interest/job satisfaction as key variables Makes 20-30 year career decisions based on 7-10 minute interview with Navy Classifier

    3. 3 Aptitude and Interest: RIDE and JOIN RIDE produces a limited set of ratings a person is qualified for and which have saleable quotas. The first tool we have developed is called RIDE or Rating Identification Engine. RIDE is replacing the longtime tool used for matching applicants to jobs for the classifier to use in selling a contract From the broad set of Navy jobs, RIDE identifies those where the applicant would have the best chance at passing the training the first time, that is, without setbacks (first term pipeline success) BLUE CIRCLE RIDE does a lot of things better than the old method: In comparative test with historical data, we’ve found that the old method made as many as 40% misclassifications and RIDE offered better matches RIDE will soon add an interest component called JOIN…a Navy-specific interest inventory..when JOIN is introduced it will help us even better define those jobs where an applicant will be successful and satisfied WHITE CIRCLE We’re currently doing T&E for RIDE at San Diego MEPS…great classifier acceptance…over 96% of applicants report they believe they got best job…but proof will be over time…does RIDE reduce attrition? The first tool we have developed is called RIDE or Rating Identification Engine. RIDE is replacing the longtime tool used for matching applicants to jobs for the classifier to use in selling a contract From the broad set of Navy jobs, RIDE identifies those where the applicant would have the best chance at passing the training the first time, that is, without setbacks (first term pipeline success) BLUE CIRCLE RIDE does a lot of things better than the old method: In comparative test with historical data, we’ve found that the old method made as many as 40% misclassifications and RIDE offered better matches RIDE will soon add an interest component called JOIN…a Navy-specific interest inventory..when JOIN is introduced it will help us even better define those jobs where an applicant will be successful and satisfied WHITE CIRCLE We’re currently doing T&E for RIDE at San Diego MEPS…great classifier acceptance…over 96% of applicants report they believe they got best job…but proof will be over time…does RIDE reduce attrition?

    4. 4 Why JOIN? Jobs & Occupational Interests in the Navy

    5. 5 Previous Research Efforts Lightfoot, McBride, Heggestad, Alley, Harmon & Rounds (1999) Existing civilian and military approaches were considered and rejected Failed to discriminate between all Navy Jobs (i.e., ratings) Scoring procedures inappropriate for classification algorithms Lightfoot, Alley, Schultz, Heggestad, & Watson (2000) Traditional interest measure developed Items not easily adaptable to future changes in Navy jobs Comprised of over 600 critical activity statements Watson (2001) The JOIN model was developed such that: An interest composite can be determined for a rating based on an SME evaluation of the job Such a model eliminates the need for conducting validation studies each time a rating is added or modified

    6. 6 Model Development Job description documents were collected for 79 Navy jobs; job descriptors extracted into 4 areas: Environment/style (e.g., indoor, outdoor, mental, physical) Community (e.g., Surface, Construction, Aviation) Process (verb; e.g., maintain, make, train) Content (noun; e.g., facilities, electronics, documents) Job description composites were created for each rating from initial analysis Process and Content conjoined to natural pairs; e.g., maintain-mechanical equipment, operate-weapons, analyze-data, etc. Composites and areas revised through iterative SME interviews Pictures collected concurrently with SME input to accurately depict examples from all areas

    7. 7 Subject Matter Expert Model 7 Navy Community Areas: a) aviation, b) construction, c) health care, d) intelligence, e) submarine, f) surface, and g) special programs 4 Work Styles & 4 Work Environments: a) indoor, b) outdoor, c) industrial, d) office, e) mental, f) physical, g) work independently, and h) teamwork 26 Work Activities or Process-Content (PC) pairs: analyze-communications, -data, -documents direct-aircraft, -emergency response maintain-documents, -electrical equipment, -electronic equipment, -facilities, mechanical equipment, -security, -supplies, -weapons make-communications, -documents, -facilities, -mechanical equipment operate-electrical equipment, -electronic equipment, -facilities, -mechanical equipment, -office equipment, -weapons respond-to emergencies, serve-customers, train-people

    8. 8 Subject Matter Expert Model

    9. 9

    10. 10

    11. 11

    12. 12

    13. 13 Preliminary Analysis - Quantitative Adequate internal consistency for individual item “pairs” average ? = 0.87; min ? = 0.72, max ? = 0.92 Initial descriptive statistics indicate that participants provided differential responses both across and within work activities Average testing time was 13 min (sd=4.5) Average time per item 8-10 sec (21 sec for SEALS screen) Special Programs and Aviation are preferred communities Working outside and in a team are preferred work style/environment Operate weapons was most preferred work activity

    14. 14 Preliminary Analysis - Qualitative Usability analysis was favorable People liked the interface People felt that the pictures gave a better impression of what jobs entailed than verbal information alone Some suggestions were made Redundancy of items Tutorials were unnecessary Their feedback served as basis for software modifications Slider bar modified to restrict erroneous responses Items ordered in blocks to reduce feeling of repetitiveness

    15. 15 Future Directions Collect data at RTC Great Lakes (~ 5,000) for analysis of vocational preference structure and interface design Validate SME model Using independent groups (e.g., “detailers”, instructors) Additional confirmatory statistical analyses Track Sailors for criterion-related validation Attrition and retention Technical training success Performance evaluations and career success factors

More Related