Friday, December 7, 2012

How Many Entry-level Analysts Will The US IC Need In 2013? (Survey)

Good question, right? 
 
If you have direct knowledge of information that might help answer the question in the title or you have indirect knowledge that is relevant to the answer to the question in the title, please take 2 minutes to complete this survey. 
 
What do I mean by direct and indirect knowledge?
Direct knowledge means that you know personally or have good information concerning the hiring plans of your agency or organization (or at least your section or division).  You might work in HR or be a manager with hiring responsibilities. 
Indirect knowledge is information that is relevant to the question that is not due to your direct responsibilities.  You might have spoken with an HR manager or have been involved in meetings where this issue was discussed. 
We are NOT looking for opinion based on purely circumstantial information.  If you are not involved in the hiring process either directly or indirectly, please DO NOT take this survey.

Why are we interested?

Every year, other disciplines announce hiring projections for the year:  "This year's hot jobs are for engineers and chimney sweeps."  That sort of thing.  Entry level intelligence analysts who are searching for a job, on the other hand, receive no such guidance.

We hope to change that.  Working with one of our hot-shot grad students, Greg Marchwinski, we put together this survey to get a better feel for the the job market for entry level analysts for the year ahead.

Once we get enough survey data, Greg will compile it and combine it with the macro-level, mostly qualitative data that we already have and put together a "jobs report" for the year ahead.  I will publish it here once we are done.

We understand that there are some legitimate security concerns here so we have tried to frame the questions such that they are focused on broad developments and general trends.  We are not interested in the kind of deep details that might compromise security.

Finally, we intend to follow this study up with similar surveys of the law enforcement and business job markets for entry-level intelligence analysts as well.

By the way, this is the same question we asked last year and here is the answer we got...

Thanks for your participation!

Monday, November 19, 2012

Want To Do Something Different This Thanksgiving?

Some of the action on DAGGRE.org
Eating, shopping, driving, yelling, more eating...I get it.  What is Thanksgiving without its traditional activities?

But, if you were looking to do something a bit different, you might want to check out the recent updates to the DAGGRE prediction market.

(I know, I know, I am such a shill for DAGGRE...)

Yes, in the interests of full disclosure, I am a member of this IARPA funded research project and yes, I probably love it way too much but the most recent changes to our site's software are, frankly...cool.

Real cool.  Like never before seen cool.

Why?  Because now you can do real-world, real time linchpin and driver analysis in a prediction market.

I am pretty excited about this but it might not be obvious why, so let me break it down.  Prediction markets typically work by asking resolvable questions like "Will Despot X be out of power by 31 DEC 2012?"  Then various people interested in that question will go into the market and make "edits" that assess the probability of the answer to this question being yes or no.  For example, if I thought Despot X would almost certainly be out of power by the end of the year, I would go in and change the probability to, say, 90%.  Someone else might go in after me and think, "What an idiot!" and change it back to 25%.  When 31 DEC rolls around one of us will be closer to right and the other closer to wrong.  The one closest to right scores the most points on the market.

This system works pretty well with straightforward questions that obviously lean strongly in one direction or another (EX:  "Will a shooting war break out between the US and Canada before 31 DEC 2012?"  Uh...no.).  It works significantly less well with more nuanced questions that really deserve to be teased apart.

This is essentially what the CIA's Deputy Director For Intelligence, Douglas MacEachin, was trying to do back in the early 1990's when he insisted on having his analysts identify linchpins and drivers.  To quote an early version of the CIA Tradecraft Manual, drivers are "key variables...that analysts judge most likely to determine the outcome of a complex situation" while linchpins are "the premises that hold the argument together and warrant the validity of the conclusion." 

This kind of analysis is pretty sophisticated stuff and it really is what makes the difference between a bald estimate ("X is 80% likely to happen") and the kind of estimate most decisionmakers expect ("Despite A and due primarily to B and C, X is 80% likely to happen.").

Before, you simply couldn't do this kind of stuff in a prediction market.  Now, with the most recent upgrade to the DAGGRE market, you can.  Let's go back to our Despot X example.  We know that rebel forces are trying to take a key city in this despot's country.  We also strongly believe that if the rebels take the city the despot's days are numbered.  Before, our estimate of the despot's longevity was a mish-mash of many factors, one of which was the possible(?)/probable(?) fall of the key city and our estimate probably was much more wishy-washy than we wanted it to be.

The DAGGRE market now lets you not only make estimates on both how likely the city is to fall and on how likely the despot is to stay in power but also allows you to make the answer to one of these questions, an assumption for the other (EX:  "Assuming key city falls, Despot X is 90% likely to be out of power by 31 DEC 2012.").

In my mind this is a huge step forward in making prediction markets more useful to real-world analysts working on real-world questions.  Definitely worth taking a few minutes to check out over Thanksgiving!

Sunday, November 11, 2012

An Interesting Perspective On What It Means To Be A Vet

If you have a few minutes this Veterans' Day, this video is worth your time...


Monday, November 5, 2012

What Can Intelligence Expect From Prediction Markets?

Opening screen of the DAGGRE.org prediction market
Prediction markets have long been touted as tools that have a wide variety of potential uses for intelligence professionals.  Far more accurate in many cases than expert judgement alone, these markets tend to incentivize good thinking and punish poor thinking in ways that, over time, produce quantifiably better results on topics like elections and sales forecasts.  Strong advocates of this method have even suggested that these markets might be able to replace traditional analysts entirely.

Naysayers have (and will likely continue...) to argue that the kinds of questions asked of intelligence professionals do not lend themselves to pat, numerical estimates.  Furthermore, they will say, even in the handful of cases where such answers would be of potential use, combining the estimates of people who know little to nothing about the details of a particular, narrow problem -- the kind that is usually of intelligence interest -- will only serve to create an estimate that is also of little to no use.  Finally, while these estimates are useless in forecasting the future (or so the naysayers will say), they will serve to anchor both intelligence professionals and policymakers alike, reducing their ability to see alternatives to the predicted outcome.

The purpose of this series of posts, then, is to explore both sides of this argument, to look at prediction markets from the point of view of the intelligence profession and in light of ongoing research and to come to some preliminary conclusions about the future of prediction markets as a tool for the working intelligence professional and the decisionmakers they support.

Informing this series will be the results of research done by the DAGGRE prediction market and the scientists involved in that effort.  DAGGRE is run by Dr. Charles Twardy of the C4I Center at George Mason University and is working in cooperation with a number of other universities and organizations (including Mercyhurst University and yours truly) to better understand prediction markets and their potential uses to the intelligence community.  

The DAGGRE project is one of five such projects funded by the Intelligence Advanced Research Projects Activity (IARPA) under their Aggregative Contingent Estimation (ACE) program.  Now in its second year, ACE has already produced a number of interesting results and promises to produce many more.  

In short, whatever your initial reaction is to the idea of prediction markets in intelligence, this series is designed to give the working intelligence professional inside access to some of the most interesting and intriguing results from research currently being done on prediction markets and intelligence questions.  My goal is to turn these results into “plain English” so that you can have an informed opinion about these unique tools.

Next Week:  What Is A Prediction Market?

Wednesday, October 24, 2012

The New HUMINT?

A few months ago, I wrote an article on the Top 5 Things Only Spies Used To Do (But Everyone Does Now).  In that article I stated that one of those things (the #2 thing, in fact) was to "run an agent network."

I equated our now everyday activity of finding and following various people on LinkedIn or Twitter to the more traditional case officer activities of spotting, vetting, recruiting and tasking agents.

While I meant that article to be a bit lighthearted, over the last several months I have been exploring this idea with some seriousness in a class I am teaching with my colleague, Cathy Pedler, and a group of very bright grad students.



The picture above gives you an inkling of the progress we have made.

In this class (called Collaborative Intelligence - "How to work in a group while learning how groups work"), we have focused our energies on critical and strategic minerals.  I have already written about this course (if you want more details go here), but suffice it to say that, recently, we decided to use our new-found skills in social network analysis to see if we could solve a traditional HUMINT problem:  "Who should we recruit next?"

Every case officer knows that their agents' value are not only measured in terms of what they know but also in terms of who they know.  Low level agents with an extensive network of contacts within a targeted area of interest are obviously valuable, perhaps even more valuable than the recluse with deep subject matter expertise.

Complicating the case officer's task, however, is the jack-of-all-trades nature of the traditional HUMINT collector.   Today, the collector needs to tap into his or her agent network to get economic information; tomorrow, political insights; the next day the need is for information to support some military or technological analysis.

Only an expert case officer with deep contacts can hope to be able to respond to the wide variety of requests for information.  In today's fast moving, crisis-of-the-day type world, the question becomes "Where can I find good sources of information ... on this particular topic ... quickly?"

Twitter to the rescue!

You see, the image I referred to earlier began as the 11 lists of Twitter users the 11 students in my class were currently following as they studied critical and strategic minerals.  The students had found these Twitter users the old fashioned way - they bumped into them.  That is, they found them on blogs or in news articles that talked about strategic mineral issues and they followed them on Twitter in order to stay current on their postings.  Since each of the students has a slightly different portfolio (the students are broken into three teams, national security, business and law enforcement and then, within those teams, each student has an area of specific interest), their lists have some common sources but many different ones as well.

The natural next question is, "Who are my sources of information following?"  Using NodeXL to collect the data and ORA to merge, manage and visualize it, the students rapidly discovered who their "agents" were following.  Furthermore, we were able to discover new people to follow -- Twitter users that many people on our initial lists were following (implying that they were potentially very good sources of information) but that the students had not yet run across in their research.

The picture got even more interesting when we merged the results from each of the students.  Once we cleaned up the resulting picture (eliminated pendant nodes, color coded the remaining Twitter users by team, etc), the students had identified over 50 new sources of information -- Twitter users who were posting information relevant to the issue of strategic minerals and vetted by many of the Twitter users we had already identified -- that we had never heard of.  You can see this more focused set of Twitter users in the image below.



While this sounds exciting (and it was, it was...), trying to listen to over 50 new voices seemed to be a bit overwhelming.  The question then became, "Of these 50, which are the 'best'?"

The traditional answer involves following all of them and then, over time, sorting out the wheat from the chaff.  Most people don't have that kind of time; we certainly didn't.  We needed another way to sort them and, thankfully, Twitter itself provides some potentially useful answers.

The first answer, of course, is to look at the number of "followers".  This is the number of Twitter accounts that claim to follow a particular person or organization.  In general, then, the sheer number of people who are following a particular person is a rough measurement of their influence and, by consequence, importance to a conversation on a particular topic.

Most twitterati don't put much credence in gross tallies of followers, though.  Anyone with a twitter account knows that only a relatively small number of their followers are actively engaged with the medium.  Some studies have also indicated that a third or more of these followers are fake or, even worse, bought and paid for.  While this is typically true on some of the most widely followed accounts and is significantly less likely to be true among the people who are tweeting about rare earths, for example, it is still a cause for concern.

Twitter again offers a solution to this problem but it takes a little work to get it.  The key is Twitter's List feature.  Twitter allows users to create lists of people; subsets, if you will, of the larger group of people a particular user might follow.  For example, I have a list of competitive intelligence librarians (there are actually quite a few on Twitter).  Lists are a way for people to follow hundreds or thousands of people but narrow and focus that chorus in a way that is most useful for them.  It allows the savvy Twitter user to filter signal from noise.

Twitter allows a user to not only look at their own lists but to know how many lists other people have created with their name on it.  This is important because it takes time and effort to create and curate a list.  It is almost certain that you have not been placed casually on a list.  Being placed on a list is an indicator of credibility; being on lots of lists even more so.  Like followers, though, the number of lists is still pretty rough and does not give the best sense of the value of a particular Twitter user to his or her followers.  Thus, while the number of lists you are on is not a bad indicator, many people like to use the list-to-follower ratio to assess overall credibility.

In other words, if you had 1000 followers and every one of them had placed you on a list, you would have a list-to-follower ratio of 1.  If only 500 had placed you on a list, then your list-to-follower ratio would be .5.  In practice, list-to-follower ratios of .1 are rare.  Based on my experience a list to follower ratio of .05 is very good and a list to follower ratio of .03 or lower is more typical.

While I am certain that there are automatic ways to collect the data you need from Twitter, we simply crowdsourced the problem.  Dividing the list into 11 pieces, we were able to quickly and accurately collect and deconflict the various data we needed including number of lists and number of followers.  In the end, we were able to rank order the 50 top Twitter users talking about Strategic Minerals in a variety of useful ways.  In all, including the teaching, it took us only about 6 hours to get from start to Top 50 list (For the complete list and more details go here)..

And here is where the analogy breaks down...

Up to this point, we were able to fairly confidently connect traditional HUMINT ideas and activities with what we were doing, much more quickly, using Twitter data.  The analogy wasn't perfect but it seemed good enough until we put the students -- the "case officers" -- into the network.  They stuck out like sore thumbs!

Case officers in traditional HUMINT networks need to be working from the shadows, pulling the strings on their networks in ways that can't be seen or easily detected.  Trying to lurk on Twitter in this sense just doesn't work, however.  My students, who are following many people but are not followed by many, became very obvious as soon as they were added to the network.  The same technology that allowed us to rapidly and efficiently come up with a pretty good first cut at who to follow on Twitter with respect to strategic minerals, allows those same people to spot the spammers and the autofollow bots and the lurkers and even the "case officers" pretty easily.

Back in my Army days we used to say, "If you can be seen you can be hit.  If you can be hit, you can be killed."  Social media appears to turn that dictum on its head: If you can't interact, you can be spotted.  If you can be spotted, you can be blocked.

It turns out, it seems, that the only way to be hidden on Twitter is to be part of the conversation.