Skip to content

SIOP 2017: Twitter Analytics and Dataset

2017 May 24
tags: ,
by Richard N. Landers

It’s been about a month since SIOP 2017. I had a whole series of post-SIOP posts planned, but due to a death in the family, my schedule’s gone upside down! But never fear, dear reader, because I’m going to post them anyway!

All of these posts will center on my major takeaway from SIOP 2017: change is afoot.  One of the ways I noticed it was the increasingly rapid shift towards technology in the conference program. When I joined SIOP over a decade ago, there were maybe one or two sessions that had anything to do with technology research. These days, you cannot escape it; there’s a significant awareness that if we don’t engage technology head on, we’re going to be left behind.

One prominent outlet for technology engagement is Twitter. When I first started paying attention to Twitter during SIOP in 2011, there was not enough activity to write about.  This year, the #siop17 hashtag had 1631 posts between April 24 and May 2, which is the SIOP conference plus a few days around the outside.  A total of 282 unique Twitter accounts used the hashtag.  Considering the importance of SIOP, I feel confident saying that this captures most Twitter users in the I/O psychology Twitterverse.  With roughly 8000 SIOP members, I would hazard a guess that this means about 3.5% of SIOP is on Twitter (and maybe 7% of conference attendees).  Not big, but not tiny, and catching up to the US numbers (about 7% of the US population is on Twitter). So what better way to celebrate the trend toward embracing technology than some Twitter analytics?

If you’d like to play with this dataset yourself, feel free to download it here. I’ll be putting together some visualizations…eventually…so if you’d rather beat me to it and show off your R or Tableau skills, feel free.

For now, here are some high-level summary tables.

Most Active #siop17 Users

  1. (146) @SIOPtweets
  2. (75) @ghiovanni_m
  3. (59) @IBMWatsonTalent
  4. (52) @HugoMunsterberg
  5. (50) @TedHayes007
  6. (39) @DrCattell
  7. (34) @LewisGarrad
  8. (32) @allietumminia
  9. (32) @lisa_s_moore
  10. (28) @HoganAssessment

Most Retweeted Tweets

  1. (14) @SIOPtweets: #SIOP17 We have a new record! 4638 registrants so far- Welcome to Orlando!
  2. (14) @LewisGarrad: Great Teams Are About Personalities, Not Just Skills #SIOP17
  3. (13) @DDIworld: Congratulations, @EvanSinar for being saluted as one of the @SIOPtweets 2017 Fellows! #SIOP17
  4. (11) @AimeeGthePHD: GRIT-same wine, new bottle? Compelling data showing construct overlap. Are we really just measuring conscientiousness and ambition? #SIOP17”
  5. (11) @EvanSinar: Filing away my #SIOP17 program – tho rarely opened or carried with me onsite thanks to Whova’s awesomeness! Great work @DrZackHorn and team!
  6. (10) @TRPoeppelman: Congratulations @EvanSinar for being awarded Fellow at #SIOP17 – Exceptionally well deserved. #prestigiousaward #soproud
  7. (10) @BoArmstrong: Let’s go ahead and start off on the right foot. It’s #SIOP17, not #SIOP2017. Follow @SIOPtweets!
  8. (9) @ALOstasis: Congrats to @stevenrogelberg on receiving the inaugural SIOP Humanitarian Award! #SIOP17 #IOPsych
  9. (9) @rnlanders: #SIOP President Mort McPhail: Replicability, oversaturation of theory, and dissemination are the key existential issues facing us #siop17
  10. (9) @PeterReiley: #SIOP17: Come IGNITE solutions to help #military #veterans transition into the workforce | @SIOPtweets #IOPsych |

Most Favorited Tweets

  1. (45) @SIOPtweets: #SIOP17 We have a new record! 4638 registrants so far- Welcome to Orlando!
  2. (37) @HugoMunsterberg: To the #SIOP17 newbie in the next hotel room: You’ve practiced enough. It’s not the Gettysburg Address you’ll be mumbling through tomorrow.
  3. (34) @MikeZickar: At the #SIOP17 session on social media, hoping that someone will stand up and say, “I am @HugoMunsterberg!”
  4. (34) @rnlanders: #SIOP17 closing plenary with @NASA has convinced me that sometimes job performance measures can be dichotomous #didntcrashintoaplanet
  5. (33) @Aaron_Kraus: Session Update: Generations are more similar than different and everyone got a trophy. #IOPsych #SIOP17
  6. (25) @DDIworld: Congratulations, @EvanSinar for being saluted as one of the @SIOPtweets 2017 Fellows! #SIOP17
  7. (24) @VeronicaRabelo_: Thank you to all the custodians, housekeepers, chefs, + other service workers who make #SIOP17 possible!! You the real MVPs!! #SIOP2017
  8. (23) @ALOstasis: Cheers to a great conference. Eagerly awaiting next year in Chicago! #SIOP17
  9. (23) @BoArmstrong: If I fail my comps tomorrow, I might just stay in Orlando as a park attendant. #SIOP17
  10. (22) @chris_cerasoli: Congratulations to @EvanSinar on being awarded Fellow at #SIOP17 @SIOPtweets

Tweets with the Most Replies

  1. (6) @rnlanders: “IO is harder than rocket science.” Wow. Let’s see if we can pat our ourselves on the back even harder. #siop17
  2. (5) @DrCattell: Considering doing an informal “best poster award” from me & @HugoMunsterberg @ #SIOP17 Hugo, you down? Bless us with your esteemed criteria!
  3. (4) @KatherineSliter: Celebrated Friday night at #SIOP17 by getting vomited on by my baby. Four times. #parenting #nofunforme #lifewithkids
  4. (3) @MikeZickar: At the #SIOP17 session on social media, hoping that someone will stand up and say, “I am @HugoMunsterberg!”
  5. (3) @BoArmstrong: If I fail my comps tomorrow, I might just stay in Orlando as a park attendant. #SIOP17
  6. (3) @MikeZickar: Loved all the energy at #SIOP17 but good to have a little work recovery time. I hope you all can get a nap on the couch before Mon hits!
  7. (3) @IOSyIslam: Some last minute #SIOP17 preparation b4 my flight tomorrow Looking 4ward 2 seeing friends & colleagues at the premier #IOpsych conference!
  8. (3) @mchetta_phd: Why don’t we do a social media panel in on the panel that @surveyguy2 @TRPoeppelman @EvanSinar and others did at #SIOP17 ?
  9. (3) @HugoMunsterberg: When’s the best time to walk unannounced into “”100 Yrs of I-O Research”” to receive my unsolicited laurels? Beginning, middle or end? #SIOP17
  10. (3) @RMendelsonPhD: Already saw #TwitterFamous peeps at #SIOP17. @surveyguy2, @mchetta_phd, @MikeZickar, and @dwbracken. Where are @Aaron_Kraus and @IOSyIslam?
  11. (3) @allietumminia: What features of a person’s LinkedIn account make them more hireable? 1) Profile length 2) Profile picture & 3) # of connections #SIOP17
  12. (3) @BoArmstrong: Creativity = Fluency, Flexibility, & Originality, but is Originality not a circular definition of creativity? #SIOP17 Please explain.
  13. (3) @DrCattell: I’m broadening to all formats: So far we have: 1) “”Too many cooks”” 2) “”Redundant Again”” award 3) “”Worst joke in Title”” What else? #SIOP17″

SIOP 2017: Schedule Planning for IO Psychology Technology

2017 April 18
by Richard N. Landers

Every year, I post some screenshots of my insane schedule for seeing all of the technology-related content at SIOP.  Well, it’s finally happened. I can’t. There’s too much. In the past, I might miss a couple of double-booked sessions, but this year, two things happened:

  1. The technology program itself ballooned which means overlaps are everywhere. You can read more about its composition in this TIP article, but in short, there are about 40 sessions or posters on the program that specified technology as one of their core categories. This doesn’t even include people who didn’t set their category to technology despite talking heavily about technology.  Frankly, within the categories I identify below – gaming & gamification, UIT & mobile testing, adapting I/O to future technology, and methodology & measurement – you could have a quite full conference only attending presentations within that category!
  2. I am personally involved in 13 different events this year, so I barely have time to attend my own presentations!

So without further ado, here is a list of things to see and when/where to see them.

Note 1. Bolded sessions are ones my lab or I am involved in.

Note 2. Occasionally you’ll see a session marked (MP).  Almost all tech sessions were listed in the SIOP program as “mixed audience,” and only one (a poster in the Innovation session 123) was listed “mostly academic.”  The ones marked “mostly practitioner” are marked here as “MP.”

Note 3. If you want to meet up with me for any reason and haven’t scheduled a time yet, my schedule is almost booked solid, but I will be wandering around the Technology and Innovation Poster Session at 8:30AM on Friday, and I will probably get there a little early. My lab and I have four posters in the session. Drop by!

Technology Sessions at SIOP 2017, April 27-29 in Orlando, FLGaming and Gamification

  1. Thursday
    1. 53: Classroom Gamification: The Impact of Gamified Quizzes on Student Learning
      1:30PM, 50 minutes, Atlantic BC
    2. 96: Gaming and Gamification IGNITE: Current Trends in Research an Application
      5:00PM, 50 minutes, N. Hemisphere E2
  2. Friday
    1. 123: Innovation and Technology Poster Session
      8:30AM, 50 minutes, Atlantic BC
    2. 135: Game-based Assessment: Concepts and Insight from Research and Practice
      10:00AM, 80 minutes, N. Hemisphere E2
    3. 153: Serious Assessment Games and Gamified Assessment: Emerging Evidence
      11:30AM, 80 minutes, N. Hemisphere A3

UIT and Mobile Testing

  1. Thursday
    1. 45-19: Mobile Versus Desktop Assessments: Is There Really a Difference? (Poster)
      12:30PM, 50 minutes, Atlantic BC
    2. 45-34: UIT Device-Type Score Differences: The Role of Working Memory (Poster)
      12:30PM, 50 minutes, Atlantic BC
    3. 87-18: Swipe Right on Personality: A Mobile Response Latency Measure (Poster)
      4;30PM, 50 minutes, Atlantic BC
  2. Friday
    1. 110: Friday Seminar (separate fee): The Use of Mobile Devices in Employment-Related Testing and Assessment
      8:00AM, 180 minutes, N. Hemisphere A1
    2. 123: Innovation and Technology Poster Session
      8:30AM, 50 minutes, Atlantic BC
    3. 130: Painting the Picture: What is the Mobile Test Environment? (Poster)
      10:00AM, 50 minutes, Atlantic BC
    4. 173: Mobile Testing “In the Wild”: Apps, Reactions, Images, Criterion Validity
      1:00PM, 80 minutes, Australia 3
  3. Saturday
    1. 244: Assessments on Mobile Devices: Our Opportunities at Digital Speed (MP)
      8:00AM, 80 minutes, N. Hemisphere A4
    2. 279: Expanding Knowledge About Mobile Assessments Across Devices and Applicants
      11:30AM, 50 minutes, Asia 5
    3. 304: Mobile Assessment: Small Screens Become Mainstream (Demo & Panel Discussion) (MP)
      12:30PM, 50 minutes, N. Hemisphere E1

Adapting I/O to Future Technology

  1. Thursday
    1. 22: The Co-Bosts Are coming! Is I-O Ready?
      10:30AM, 50 minutes, S. Hemisphere V
    2. 36: Workplace Automation & the Future of IO Psychology
      12:00PM, 80 minutes, N. Hemisphere A4
    3. 60: Symposium + Panel Combo: Will Technology Make Assessment Obsolete?
      1:30PM, 80 minutes, N. Hemisphere E2
    4. 70: Fostering Collaboration Between Data/Computer Scientists and I-Os
      3:30PM, 50 minutes, Asia 3
  2. Friday
    1. 122: Science-Practice Exchange: Ready or not..Technology’s Implications for Leadership Development
      8:00AM, 80 minutes, S. Hemisphere V
    2. 123: Innovation and Technology Poster Session
      8:30AM, 50 minutes, Atlantic BC
    3. 167: From Likes to Impact: The Payoffs of Social Media Involvement (MP)
      1:00PM, 80 minutes, Asia 1
    4. 196: Opportunities and Challenges in Electronic HRM
      3:00PM, 80 minutes, N. Hemisphere E1
    5. 207: Technology Trends Leading HR Practice: Key Opportunities for Research?
      400PM, 50 minutes, Asia 3
    6. 223: Executive Board Special Session: Simple, Social SIOP: Collaborating to Increase SIOP’s Social Media Reach
      5:00PM, 50 minutes, Asia 2
  3. Saturday
    1. 291: Integrations and Partnering with Technology: Experiences and Best Practices (MP)
      11:30AM, 80 minutes, S. Hemisphere IV
    2. 309: From the Outside, In: Technology’s Influence on I-O Psychology (MP)
      12:30PM, 50 minutes, S. Hemisphere II
    3. 311: The I-O of the Future: Identifying and Closing Skills Gaps
      12:30PM, 50 minutes, S. Hemisphere V
    4. 328: Alliance Special Session: The Impact of Technology on Recruitment and Selection: International Perspectives
      1:30PM, 50 minutes, N. Hemisphere E4
    5. 329: I-O Psychology in an IT World (MP)
      1:30PM, 50 minutes, S. Hemisphere I

Methodology and Measurement

  1. Thursday
    1. 25: Panel + Breakout Combo Session: Sense Making of Wearable Sensors
      11:30AM, 50 minutes, S. Hemisphere IV
    2. 44: Community of Interest: Inductive Research in I-O Psychology
      12:30PM, 50 minutes, Asia 3
    3. 53: Insufficient Effort Responding in MTurk Research: Evidence-Based Quality Control (Poster)
      1:30PM, 50 minutes, Atlantic BC
    4. 53: Comparing MTurk and the U.S. Populations’ Occupational Diversity (Poster)
      1:30PM, 50 minutes, Atlantic BC
    5. 53: Evaluating Online Data Quality: Response Speed and Response Consistency (Poster)
      1:30PM, 50 minutes, Atlantic BC
    6. 157: Master Tutorial: Automated Data Collection: An Introduction to Web Scraping with Python
      11:30AM, 80 minutes, N. Hemisphere E3
  2. Friday
    1. 123: Innovation and Technology Poster Session
      8:30AM, 50 minutes, Atlantic BC
    2. 130: Cheating on Online Cognitive Tests: Prevalence and Impact on Validity
      10:00AM, 50 minutes, Atlantic BC
    3. 193: Friday Seminar (separate fee): Automated Conversion of Social Media into Data: Demonstration and Tutorial
      3:00PM, 180 minutes, N. Hemisphere A2
  3. Saturday
    1. 239: I See What You Did there: Data Visualization in Action
      8:00AM, 80 minutes, S. Hemisphere I
    2. 259: Master Tutorial: Making Research Reproducible: Tutorial for Reproducible Research with R Markdown
      10:00AM, 80 minutes, Asia 4
    3. 265: Master Tutorial: R Shiny: Using Apps to Support I-O Research
      10:00AM, 80 minutes, N. Hemisphere A3
    4. 301: MTurk as Work (and Not Just a Recruitment Method)
      12:30PM, 50 minutes, N. Hemisphere A2
    5. 336: Using New metaBUS Functions to Facilitate Systematic Reviews and Meta-Analyses
      3:00PM, 80 minutes, Asia 4
    6. 345: Data Visualization with R
      3:00PM, 80 minutes, S. Hemisphere I

Special/Other Topics

  1. 24-19: Examining the Relationship Between Engagement and Technology-Assisted Supplemental Work (Poster)
    Thursday, 11:30AM, 50 minutes, Atlantic BC
  2. 38: Caught on Video: Best Practices in One-Way Video Interviewing (MP)
    Thursday, 12:00PM, 80 minutes, N. Hemisphere E2
  3. 87-13: Predicting Personality with Social Media Behavior: A Meta-Analysis (Poster)
    Thursday, 4:30PM, 50 minutes, Atlantic BC
  4. 261-22: A Meta-Analysis Comparing Face-to-Face, Online, and Hybrid Ethics Courses (Poster)
    Saturday, 10:00AM, 50 minutes, Atlantic BC
  5. 268: Social Media for Employment Decisions: The Good, Bad, and Ugly
    Saturday, 10:00AM, 80 minutes, N. Hemisphere E2

Technology Poster Session

If you’d like to whet your appetite for the technology poster session, here are the titles you’ll see there (Friday, 8:30 AM, Atlantic BC):

  1. How Pay Affects Performance and Retention in Longitudinal Crowdsourced Research
  2. Aristotle, Kant, and Facebook? Implications of Social Media on Ethics
  3. Can Video Games Reduce Faking in Selection Assessments?
  4. Coworker Relationships Altered by Social Media: Posts, Pokes, and Problems
  5. Time Flies When Cognitive Tests Are Games
  6. Diversity and Group Creativity in an Online, Asynchronous Environment
  7. The Effect of Technology Use on Relationship and Network Development
  8. The Impact of Smartphone Usage on Perceptions of Work–Life Balance
  9. How Consistent Is the Impact of Devices on Working Memory?
  10. Crowdsourcing Hard-to-Reach I-O Psychology Populations: Feasibility and Psychometrics
  11. Increases in Applicant Pool Diversity Attributable to Unproctored Internet-Based Testing
  12. Creating Three-Dimensional Task–Technology Fit Scales
  13. Examination of Individual Differences in Preference in Pursuing Gamified Training
  14. Email Me! How Email Textual Cues Influence Perceptions
  15. Personality, Responsiveness, and Performance in Technology-Enabled Work Environments
  16. To Meet or Not to Meet: Preference for Electronic Communication
  17. Effects of Automated Technology on Experiences of Agency at Work

In the I-O Psychology Credibility Crisis, Talk is Cheap

2017 April 5
by Richard N. Landers

With the backdrop of science’s replication and now credibility crisis, triggered perhaps by psychology but now much broader than that, I/O psychology is finally beginning to reflect. This is great news. Deniz Ones and colleagues recently published a paper in our field’s professional journal, the Industrial-Organizational Psychologist (TIP), providing some perspective on the unique version of this crisis that I/O now faces, a crisis not only of external credibility but also of internal credibility. In short, I/O practitioners and the world more broadly increasingly do not believe that I/O academics have much useful to say or contribute to real workplaces.  Ones  and colleagues outline six indicators/causes of this crisis:

  • an overemphasis on theory
  • a proliferation of, and fixation on, trivial methodological minutiae
  • a suppression of exploration and a repression of innovation
  • an unhealthy obsession with publication while ignoring practical issues
  • a tendency to be distracted by fads
  • a growing habit of losing real-world influence to other fields.

I will let you read the description of each of these individually in the authors’ own words, but let me just say that I agree with about 95% of their comments and would even go a step further. Our field is just past the front edge of a credibility crisis, one that has been bubbling beneath the surface for the last couple of decades. Things aren’t irreversible yet, but they could be soon.  The only valid response at this point is to ask, “what should we do about it?” Ones and colleagues offer an unsatisfying recommendation:

To the old guard, we say: Be inclusive, train and mentor scientist–practitioners. Do not stand in the way of the scientist–practitioner model. To the rising new generation, we say: Learn from both science and practice, and chart your own path of practical science, discovery, and innovation. You can keep I-O psychology alive and relevant. Its future depends on you.

To be clear, I completely agree with this in principle. The training model needs to be changed. Researchers need to follow best path of investigation for their personal scientific research questions. We must integrate science and practice into a coherent whole. But if the replication crisis has taught me anything, it’s that waiting for the old guard to become levers of change is too slow and too damaging to the rest of us to accept. When charting your own path of practical science can lead to denial of tenure, the system is fundamentally broken. So if you want change, you must be the lever. I have tried to be such a lever, in my own small way, since I earned my PhD eight years ago. But I am eager for more to join me.

Few of you know this, but I named this blog “NeoAcademic” when I graduated in 2009 not only because I was a newly minted PhD about to set off into academia but also because I was already frustrated then with “the way we do things.” There seemed to be arbitrary “rules” for publishing in top journals that had relatively little to do with “making a contribution,” at least the way I understood the traditional meaning of the word, “contribution.” I heard grumblings from my adviser and the remainder of the Minnesota faculty about how publishing expectations for the Journal of Applied Psychology were “changing” and not necessarily for the better. But it wasn’t a “crisis.” Not yet. I just thought something was up, although I wasn’t quite sure what it was, at the time.

Focused on my own path to tenure, I saw technology as a pressing need for I/O psychology, so I naively decided to innovate through research, hearkening back to the I/O psychologists of old that actually discovered things useful to the world of work. That is, after all, why I wanted to go into academia – to trail blaze at the forefront of knowledge. As anyone who has done innovative research knows, and something I underappreciated at the time, the risk/reward ratio becomes quite a bit poorer when you innovate through research.  Several folks here told me I was crazy for trying such a thing, especially pre-tenure. Why would you spend two years investigating a hypothesis that might not turn out to be statistically significant? Because it’s important to discover the answer, regardless of what happens; that’s why.

Unfortunately, I soon discovered that our top journals did (and do) not particularly appreciate this sort of approach. True innovation, when you operate at the boundaries of what is known, is messy and mixes inductive and deductive elements. It involves the creation of new paradigms, new methodologies, and operating in a space in the research literature that is not well explored. As anyone who has published in Journal of Applied Psychology knows, that’s not “the formula” that will get you published there. True innovation is treated a “nice to have,” not a driving reason to conduct research. That felt backwards to me. It felt like our journals were publishing information that was of relatively dubious worth to any real, live human being other than other people trying to publish in those journals. It was definitely not a path toward building a “practical science.” So with that thought, still pre-tenure, I found myself a bit lost.

My initial reaction to this was to think that business schools would save me, that surely business schools would be focused on helping businesses and not just writing papers that other academics will read.  I even joined the executive committee of the Organizational Behavior division of the Academy of Management as a way to expose myself to that side of the “organizational sciences.”  Unfortunately, I discovered that the problem there is even worse than it is here. For example, most business schools actually name a short list of “A journals” that one must publish in to be tenured and/or promoted.  That seemed absolutely backwards to me.  It the very worst possible realization of goal setting. It stifles innovation and limits the real world impact of our work. It explicitly creates an “in club,” establishing pedigree as the most important driver of high quality publishing instead of scientific merit. And perhaps worst of all, it encourages people to fudge their results, or at least to take desirable forking paths, for fame and glory. This sort of thinking is now deeply ingrained in much of Academy; it has become cultural.

So with that observation, I decided business schools were a cure worse than the disease, at least for me. I would find little practical, useful science there, at least in OB.  (To be fair, when I raised this issue with an HR colleague, they just rolled their eyes at me and said, “Yeah, well, that’s OB.”)  In response to this discovery, I turned to colleagues in disciplines outside of both psychology and the organizational sciences.   Still later, I turned to interdisciplinary publication outlets. My eyes were opened.

In short, if all you know is I/O psychology, you have an incredibly narrow and limited view of both academic research and the organizational context in which we attempt to operate. I/O psychology is tiny. It often doesn’t feel that way when you’re sitting at a plenary among the thousands at SIOP, or even at Academy, but the ratio of I/O psychologists to the number of people practicing HR related tasks in the world related to what we do is essentially zero. We are a small voice in one corner of one room of a very large building. I/O psychologists often complain that we don’t get a “seat at the big table,” a voice on the world or even national stage, but that’s essentially impossible to do when we won’t let anyone anyone sit at ours either by imposing strange and arbitrary rules on “the way things are done.”  We’ve become a clique, and that’s dangerous.

Just like business schools, I/O psychologists use an intense focus on methodological rigor and “rigorous” theory as a way to exclude people who don’t have the “right” training. I recall vividly a time when I once described the expectations for publication in I/O journals to a computer science colleague. I told him offhandedly that a paper could be rejected because its scales were not sufficiently validated in prior studies. He literally laughed in my face. What about situations where previously validated scales were unavailable, or perhaps a situation where scale length was a concern? What if the scale just didn’t work out like you’d expected, despite good reason to trust it? What if data were available in a context where they usually were not? Apparently no study at all is better than one with an alpha = .6.

Now, the bright side. Unlike business schools, I/O psychology now graduates a huge number of Master’s and Ph.D. practitioners who see academia’s sleight of hand for what it really is and then call us out on it. Yet what many academic I/Os still don’t realize, I think, is that many of our students go into practice not because they can’t “hack it” in academia, but because they see the hoops and hurdles to getting anything published, which is presented as the sole coin of the realm, as unpleasant, unnecessary, and lacking or at least diminishing any impact of that research on the real world.  This growing chorus of voices, I think, is our field’s saving grace – if only we listen and change in response.

That brings me all the way back around the core question I asked at the beginning of this article: what do we do now?  The short answer is 1) change your behavior and 2) get loud. Here are some recommendations.

  1. If your paper is rejected from a journal because you purposefully and knowingly traded away some internal validity for external validity, complain loudly.  If the only way you can reasonably address a research question is by looking at correlational data inside a single convenient organization, the editor should not use criticisms of your sampling strategy as a reason to reject your paper. If Mechanical Turk is the best way, same thing. Do not let journal gatekeepers get away with applying their own personal morality to your paper. Many (although not all) of these people in positions of power are part of the “old guard” because they exploited the system themselves and expect those that follow behind them to do the same. You need to let them know that this sort of rejection is unacceptable. Do not do this with the expectation or even hope that you will get your manuscript un-rejected. Do this because it’s right. Change one mind at a time.
  2. Prioritize submissions to journals that engage in ethical editorial practices. Even better, submit to these journals exclusively.  Journals live and die by citation and attention.  If you ignore journals engaging in poor practices, they will either adapt or disappear. I’ve talked about this elsewhere on this blog, suggesting some specific journals you should pay attention to, but this list is growing. Can you imagine the effect of a journal with an 8% acceptance rate suddenly ballooning to 30% because so many people stop submitting there? Like night and day.
  3. Casually engage in the broader online community. There are many social media sites now devoted to discussing the failings of psychology, and the credibility crisis broadly, and what to do about these problems. Be a part of these discussions. Don’t let other fields, other people or the news media dominate these conversations. IO psychology is strangely quiet online, limited to Reddit, a handful of Facebook and LinkedIn groups, and an even smaller number of blogs. In the modern era, this lack of representation diminishes our influence. Speak up. You don’t need to be sharing the results of a research study to have something worthwhile to say.
  4. It’s time for some difficult self-reflection: Is your research useless? In another TIP article this issue, Alyssa Perez and colleagues went through the “top 15” journals in I/O psychology, concluding that of 955 articles published in 2016, only 47 reached standards they would consider potentially having “significant practical utility.”  Would your publications be one of the 47 or one of the 908? Importantly, Perez just developed some potential criteria for making a utility judgment, and perhaps quite stringent ones, but I would call your attention to these questions to ask yourself of each of your papers, derived from their list.  If you answer “no” to any of these, it’s time for a change. And if you think any of these aren’t worthwhile goals, you’re part of the problem:
    • Does it tackle a pressing problem in organizations, either directly or indirectly, including but not limited to SIOP’s list of Top 10 Workplace Trends?
    • Can the results of the study be applied across multiple types of workplace settings (as opposed to only advancing theoretical understanding)?
    • Are the effect sizes of the study both significant and practically meaningful?
  5. Share your research and perspective beyond the borders of I/O psychology.  Research is not the sole domain of academics; we are all scientist-practitioners, and we all have a shared responsibility to conduct research and spread it not only among each other but also among all decision-makers in the world. And if “the world” is scary, feel free to start with “HR.” Don’t hold your nose while you’re over there. Share your passion for scientific rigor and engage in good faith – because such rigor is truly a rare and significant strength of our field – but don’t use that expertise as a tool to brush aside the hoi palloi.
  6. Innovate by seeking out what I/O-adjacent fields are doing and learn all about them as if you were in graduate school again (or still are).  This is a personal pet peeve of mine; a lot of I/O psychologists, both practitioner and academic, want to operate in a bubble. They want to apply their own little sphere of knowledge, whether it’s what they learned in grad school or what they took that workshop on once, use that and be done. It’s not that easy anymore. It shouldn’t be. Self-directed learning is becoming increasingly important as the speed of innovation increases in the tech sector. Many bread-and-butter I/O psychologists can already be replaced with algorithms, and this will only get worse in the coming years. Make sure you have a skill set that can’t be automated, and if you’re at risk of being automated, go get more skill sets. I’m trying to do my part over at TIP, but this is ultimately something I/Os need to do for themselves. If you need motivation, imagine the day when an algorithm can develop, administer, and psychometrically validate its own scales and simulations. That day is coming sooner than you think.
  7. Accept that sometimes, engineering can be more important than science.  When doing science, we aim to use the experiences and circumstances of a single sample, or group of samples, to draw conclusions about the state of the world. When doing engineering, we try to use science to solve a specific problem for one population that is staring us in the face. I/O practitioners, most often, are engineering. They are using the research literature and their own judgment to create solutions to specific problems that specific organizations are facing. These are learning opportunities, all. They need to have a place in our literature and conversations. They are also the primary avenue by which fields outside of I/O learn about who we are and what we do. You must actively try to be both a scientist of psychology and an engineer of organizational solutions. And when your engineering project is successful, remember to trumpet the value of I/O psychology for getting you there.

Ultimately, I think these actions and others like them are not only the way to make academic I/O psychology relevant again but also the way to distinguish ourselves from business schools. Let them have their hundred boxes and arrows, four-way interactions, and tiny effects sizes. Leave solving the real problems to I/O psychology.