JobElephant in collaboration with our industry partners has worked to bring standardization to the way website traffic is measured for both Employers and Publishers.  

As recruitment marketing evolves toward interconnectivity and interdependence of advertisers, publishers, and technology solution providers, so grows the importance of establishing standards and transparency in how job advertising traffic quality is defined and valued.

In the absence of standards, measuring the effectiveness of job advertising media spend across publishers and analytics suites – whether purchased under duration, subscription or a pay-for-performance (CPM, CPC, CPA) payment model – is difficult if not impossible.

The problem is worth solving
1. Almost half – 48.5% – of all online traffic is not human (Imperva Incapsula Bot Traffic Report, 2015) – do bot clicks count?
2. For the U.S. market and for European markets, does foreign traffic count? For example, 16.5% of Austrians want a job outside of Austria (Indeed EU15 Movers Index, 2015), but for many employers, they would not be considered viable candidates for jobs advertised in U.S. or other European markets.

About this document
The TAtech Traffic Quality Declaration is a voluntary publisher self-evaluation. It was developed by an international Working Group established by TAtech: The Association for Talent Acquisition Solutions to bring standards and transparency to traffic measurement for the job advertising industry. The Declaration is endorsed by TAtech as its official position on the importance of establishing standards for defining and measuring traffic quality.
Employers (Advertisers) deserve
1. Transparency – understanding what traffic counts and what doesn’t is the foundation for evaluation.
2. Honest and fair reporting – reports that accurately describe the value that has been delivered.

Publishers Deserve
1. A fair playing field – empowering employers to make “apples to apples” value assessments through
2. Industry standards – a benchmark to be compared with and against that is endorsed and which can be
implemented across the industry by publishing peers.

Goal of this Declaration
This Declaration aims to build understanding and trust amongst recruitment advertising buyers and sellers by:
1. Establishing a common set of terms and terminology for discussions about traffic quality, see Traffic
Quality Determinants below.
2. Providing a framework for self-disclosure of traffic quality practices employed by publishers and
technology solution providers (analytics, ad engines, etc)
3. Leveling the ‘playing field’, to facilitate equitable measurement of all job advertising media regardless of whether the transaction is unit (duration), time (subscription) or performance (CPC, CPA)-based.
See the Education section of this document below for additional background information and useful resources for further understanding.

Traffic Quality Determinants
Technographic signals – involve inspecting the unique technical aspects of each traffic event
1. Human vs. Non-Human – is the advertising action driven by a person or a machine?
2. Geography- is the physical location of the user likely relevant based on geographic proximity to the
advertised need?

Behavioral signals – involve looking at the behavior of a given user over time
1. Duplication and Frequency – are repeated actions driving value? Is quality associated with the amount of
time between actions?

The Declaration Guidelines for Publishers
What you do vs. How you do it – In order to provide transparency without revealing information that could enable bad actors to avoid detection, describe your methods in general terms without providing vendor or tactical details. For example, say that you use software to evaluate non-human User Agent signals, but don’t name the vendor or libraries used.
Flag – To provide consistency across publisher declarations, use the term ‘flag’ to mean the process of identifying the particular activity.
Transaction Model – Unit (duration) and time-based (subscription) job advertising providers should generally answer N/A for the ‘Billable’ portion of the response. ‘Viewable on Reports’ applies to all transaction models.



Declaration Of Traffic Quality

Do you differentiate human vs. non-human advertising activity?
1. User Agent – do you analyze user agent contents?
Parsing user agent strings for declarative terms such as ‘bot’, ‘crawler’, ‘spider’, etc. and indicative terms
such as ‘phantomjs’ are useful in the identification of traffic that is not from humans.

Our technology does not count bots or spiders.
It filters them out and tracks Human based clicks. ad tracking technology is used by all advertising done through recruitment advertising agency as well as The Key Job Board Connection network of niche job boards listed here:
We additionally use Google Analytics.

2. IP Address Filters – Do you employ manual or automated traffic filtering based on the number of times that individual IP addresses drive advertising actions?
Filtering IPs that execute 100 clicks without any conversions would indicate a likelihood that the IP is not acting like a human.

We do not use any advanced IP filtering systems.

3. IP Hosts – Do you employ manual or automated filtering based on characteristics of the hosting service provider?
Certain hosts have low likelihood of being used by people who are interested in particular job advertisements. Examples include Amazon AWS, Digital Ocean, Cloud Sigma, etc.

We do not flag IP Hosts.

4. IP Geolocation – Do you employ manual or automated filtering based the IP proximity of repeated actions?
Sophisticated non-human activity can involve the use of IP Proxies that individually do not signal potential non-human activity, but when looked at as a group of actions from a single location, a more indicative pattern emerges.

Our technology does not flag traffic via IP Geolocation or Location Relevance. However we do use Google Analytics to report on locations of clicks.

5. Explicit Validation – Do you employ reCAPTCHA or other automated validation solutions such as mouse movement to confirm human activity?

We do not have control over publisher sites where we post. E.g.
On our websites and company website we run reCAPTCHA. also scans our websites for bots and spiders.

How do you distinguish traffic by geographic proximity?

6. Location Relevance – do you employ detection practices to determine geographic relevance?
The physical location of the user viewing and taking action on the ad makes a difference. For example, is the user clicking on a U.S. job listing from Russia or known to reside in another location (distant from the job posting location) based on profile/resume info?

Our technology does not flag traffic via IP Geolocation or Location Relevance. However we do use Google Analytics to report on locations of clicks.

How do you distinguish traffic by frequency and duplicity?
7. Duplicity / Frequency – do you distinguish traffic by the number of times a user interacts with an advertisement and/or how much time elapses between interactions?
It’s important to measure the number of times that a single user interacts with an ad. For example, if the same user clicks on the same job 10 times. or if a single user interacts with 10,000 ads, it’s unlikely that that is legitimate job seeking activity.
It’s also important to measure the time span between ad interactions. Is it different if a user clicks on the same job 5 times in a minute compared with once per week over the span of 5 weeks? What if a user clicks on the same ad twice in 2 seconds?

Our technology does not flag duplicity or frequency. We can flag frequency via Google Analytics.


Publishers that voluntarily complete the Traffic Quality Declaration will be upholding honest, fair and transparent standards of traffic quality reporting. Should advertiser questions or concerns arise, however, remediation can be achieved by:
1. Communication – trust and transparency can generally be achieved through open conversation between buyer and seller.
2. Policy adjustment – buyer and seller may decide to alter the terms of applicable agreements in order to more closely align interests.
3. Term adjustment – buyer and seller may decide to define the procedures for the adjustment of payment terms in the event of measurement variance outside an agreed range. Generally, a 5-10% delta between measuring systems would be considered acceptable.

Who cares?
Why is it important to pay attention to this issue even if you don’t buy/sell on a pay-for-performance model? See the following presentation deck for answers – Traffic Quality – What is it, who determines it and how?
What is the “Traffic Quality Problem”?
Internet traffic comes in all shapes and sizes, and can be measured in a variety of ways. However, in advertising specifically, Traffic Quality refers to the identification, measurement, and handling of both valid and invalid user traffic. Valid user traffic is the result of genuine user interest; a real human viewing and potentially clicking on ads; traffic we honestly feel an advertiser should pay for.
Invalid traffic, on the other hand, provides no tangible value to a legitimate advertiser.

Examples of invalid traffic include:
1. GoogleBot, a legit and well-known crawler, crawling a website to add content to the Google search index
2. An errant “double click” on a link or ad
3. A malicious click bot designed to inflate a publisher’s click volume
4. A malicious click bot designed to deplete a competing advertiser’s budget
5. And so on and so forth
Other resources on this topic include: