Bots Are Getting Good at Mimicking Engagement
Key topics
The article discusses how a significant portion of e-commerce website traffic is fake and generated by bots, and the HN discussion revolves around the implications of this on online advertising and the potential beneficiaries of such fake traffic.
Snapshot generated from the HN discussion
Discussion Activity
Very active discussionFirst comment
N/A
Peak period
99
0-6h
Avg / period
17.8
Based on 160 loaded comments
Key moments
- 01Story posted
Oct 15, 2025 at 7:11 AM EDT
3 months ago
Step 01 - 02First comment
Oct 15, 2025 at 7:11 AM EDT
0s after posting
Step 02 - 03Peak activity
99 comments in 0-6h
Hottest window of the conversation
Step 03 - 04Latest activity
Oct 17, 2025 at 2:40 PM EDT
3 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
The bots are getting creepily good at mimicking engagement. I wrote up my findings, including some of the bizarre patterns I saw and the off-the-record conversations I had with ad tech insiders. It seems like a massive, open secret that nobody wants to talk about because the whole system is propped up by it.
I'm curious if other developers, founders, or marketers here have seen similar discrepancies in their own data.
I don't click display ads or links to Amazon or YouTube listings, etc. If I want to buy something I try to go directly to the manufacture's site and search for it there. If I have to go to a third party sales platform I'll go there and search for the thing I want.
I rarely click on YouTube recommedations, because they are more and more just AI slop. I subscribe to people whose content is interesting, and that's the vast majority of what I watch.
Only considering old school "Ads by google" banners as ads and then patting yourself on the back for never clicking them is pretty faint praise for your ad evasion skills.
The thing was that site kept reflowing the layout over and over again and I think the point was that they were hoping you were going to click on something you saw on the sidebar and then it would reflow and an ad would be there right at the second when you clicked and then... Ka-Ching!
[1] Funny reversal that the enterprise product is mass market and the consumer product is overpriced if not gold-plated.
When a company puts up a billboard or an ad on the bus, they don't care if the ad is seen by dashcams and dogs. All that matters is impact on the bottom line.
Uh, yes.
If you get 47 sales on $10k in ad spend (pay per click) and $9900 of that $10k was fraudulent then you got 47 sales on $100 of ad spend. Imagine if you could stop those fraudulent clicks.
You can’t stop fraudulent clicks just like you can’t stop your SuperBowl ad from playing while your viewers are in the bathroom. How much of ESPN’s viewership happens at bars where nobody is watching?
At some point it’s not reasonable to expect ad networks to be able to stop sophisticated bots or exclude them from your billed impressions.
They should definitely try to minimize it if they want to maintain the value of their impressions but I think there is a good argument that OP just isn’t the right customer for this type of advertisement.
If you’re trying to sell a t-shirt you don’t hire a salesperson to cold call people, maybe OP shouldn’t be using web ads in the first place. If fraud was cut down by half would their situation really be that much better?
Yes, you literally can stop fraudulent clicks. I'm not sure where you get the idea that it's not possible?
This is nothing like the Super Bowl analogy.
It does, because it changes the strategy.
If you think the ads are working and have 10k potential customers then you start thinking about how to increase your conversion rate thinking you could get a chunk of those 10k, you might think distribution is solved.
But if it turns out only 2.5k are real humans then your conversion rate might not even be an issue and it’s just the marketing strategy that needs tweaking.
The whole point is that they are giving you fraudulent traffic which you use as real data to figure out the next steps. If you don’t know it’s fraudulent or how much of the clicks are fraudulent then you are taking decisions under the wrong assumptions.
> You can’t stop fraudulent clicks just like you can’t stop your SuperBowl ad from playing while your viewers are in the bathroom
That’s not even a good analogy, we are taking clicks, not impressions.
The point here, for me, is that a microeconomic perspective on this whole question is more salient than a purely technical one.
The online ad market is extremely inefficient because it has no idea how much ad spend is even reaching people.
Fake clicks give the illusion that ads are working and instead you have to optimize your funnel or whatever else.
I am fine with spending $10,000 on ads (or whatever amount).
The issue is that when I know $5,000 of it was spent on clicks that had 0% chance to convert. For every fraudulent click I can block/prevent that is one more click that can be made by a real person who may actually make a purchase.
Your calculations make as much sense as pouring out 90% of a can of beer, claiming you have just made vodka, and simultaneously trying to pay only for the 10% of beer that is left.
The entire premise of the article is quite accurate and most companies are spending money on third-parties to do what this guy has done to recoup fraudulent spends; so, none of this is new to the industry. Everyone knows it's happening and it's all part of the gig; it benefits everyone except those spending so little is truly done to correct it.
It is the ad network's problem because they are showing ads to all those bots which costs them. It is also the ad networks problem in that they are not effective - maybe for your $10k for 47 sales is good enough, but for most it isn't and so they lose customers who pay attention to value. It is also their problem in that by giving a false number of views they lose some face, but this is always a proxy that isn't of much value to anyone.
The more important thing that is your problem is verifying the ads are worth it. How do you know the ad resulted in 47 sales. A large number of adds should be buy this in the future not today, and thus clicks are the wrong measure.
You do understand that when I stop/reduce the fraudulent clicks that leaves extra ad budget available to be spent by clicks from real people who will actually buy things.
I have 0 incentive to not work to stop/reduce the fraudulent clicks. My ad spend stays the same but my sales go up.
You aren’t paying for conversion rate, you are paying for a link being put on a website when a query is made. You can’t control whether a bot follows that link. You can’t control how sophisticated that bot is. You can’t expect an advertiser to filter out every type of illegitimate traffic (although it sounds like they probably have the capability to filter out more but don’t have any incentive to do so).
I have seen recommendations from across the Internet to not bother with Google ads and other similar paid ad services. It’s basically like paying for a cold lead, you’re attracting one of the least interested types of customers.
The recommendation I’ve always seen is that it’s better to build legitimate interest in your product by producing content. Or perhaps move to an advertising platform where there’s more of a guarantee of reaching human users.
But still, I’ve heard that trying to spend customer acquisition dollars on one-time purchases is a losing battle.
If Tesla was able to start a massive car company without buying ads you can go without AdWords, too.
Google Ads used to be very effective. You are catching the punter as they are actually looking for a solution to their problem. However Google have inflated bids and increased the complexity to the point where very few people can make a decent return now.
>The recommendation I’ve always seen is that it’s better to build legitimate interest in your product by producing content.
That is becoming rapidly less true as AIs steal all the traffic.
>If Tesla was able to start a massive car company without buying ads you can go without AdWords, too.
We can't all be Teslas.
But within their advertising market, you compete for placement with other advertisers. If everyone is getting lots of fraud traffic, presumably they adjust their bids for it, if you're getting outbid consistently, it's reasonable to expect that the other advertisers are either getting a better ROI or they have a lower ROI target than you do.
About a million years ago, I was on a team that had a significant ad program, and it was primarily data driven, we'd come up with keywords to advertise on, measure the results and adjust our bids. With a little bit of manual work to filter out inappropriate keywords or to allow a lower ROI on "important" keywords. Of course, our revenue was largely also from advertising, so it was a bit circular.
Remember what matters is how the ad affects the bottom line. Everything else is just a proxy - you need to check to see if your proxies are good enough.
If you display ads on your website, then sure, it doesn't matter if a bot or real person viewed/clicked on your ad if you get paid regardless - to the point where there's a big industry of putting ads on sites then having bots engage with the site and its ads.
Billboards are a bit different because that's paid per week / month, but internet ads are paid per impression or click.
Billboards sell for flat rates, online advertising is sold on the basis of impressions.
Contrast the billboard with an online ad: If you pay for 10,000 impressions, and 9,000 of those impressions end up being automated traffic, then you’ve effectively lost 90% of your advertising budget. Like you mentioned, ad buyers can and do account for this, so it’s not clear that it’s as big of a problem as this thread is making it out to be. The theory I see thrown around is that Meta and Alphabet are tacitly allowing this bot traffic to continue (or even that they are responsible for it themselves) because it allows them to sell another round of ads, but if this is what is going on it isn’t clear why firms would keep buying ads that aren’t converting.
they also don't get charged retroactively for each dashcam and dog that looks at the billboard
But it's incredibly unlikely that the amount of fraud is stable, any attention from a small player can generate enough volume to eclipse all of the background. Thus no, you can't just account for it on your calculation.
When a company puts up a billboard, it doesn't get taken down because a flock of birds passed by before anyone could look at it.
Effective advertising depends on iterative testing, which is very hard if the signal to noise ratio is way off.
We at engineering decided to filter out bots. Figures fell dramatically by more than 50%.
In less that a day business mandated us to remove the filter.
Bots are real people after all
Interestingly, I think money is increasingly its own falsehood now. A lot of rich people are finding that they pay a lot to get what's basically a scam, like Sam Altman's swimming pool that leaked and ruined the rest of the house [1]. There's a reason that Billionaire's Bunker [2] is entering the cultural zeitgeist despite fairly terrible plotting, dialogue, and acting.
[1] https://fortune.com/2024/07/17/sam-altman-infinity-pool-mans...
[2] https://www.netflix.com/title/81606699
The correct thing to do, probably, is to just provide the new data to the customer without changing what they were already looking at. So a new widget appears on their dashboard, "52% bot traffic", they click on that, and they see their familiar line chart of "impressions over time" broken down as a stacked line chart, bottom is "human impressions over time," top is "bot impressions over time," and the percentage that they were looking at is reported either above or beneath the graph for the same time intervals. Thus calling attention to the bottom graph, "human impressions over time," and they can ask your sales people "how do I get THAT number on my normal dashboard?" and they can hem and haw about how you have to upgrade to the Extended Analytics Experience tier for us to display that info and other business nonsense...
Point is, you stimulate curiosity with loud interference rather than quietly interfering with the status quo.
If YouTube served gigabytes of the video file for 40 munutes and human watched it for that time, but they didn't send a request to `youtube.com/api/stats/atr/` and periodically to `/stats/qoe?`, did the video actually get viewed?
I think a reasonable person would say that the person viewed that video. Only a programmer would suggest that wasn't a view because they didn't properly engage the statistics and tracking endpoints.
But so much of the industry is built on deeply invasive tracking and third-party ad networks that this is a normal thing.
If that causes a problem with Youtubers making a paycheck from external sponsors, Youtube really really does not care, because sponsorships are money that youtube doesn't get!
Youtube is downright hostile to creators who don't make the "right" content, which means new videos at a perfect schedule that all have the exact same "taste" such that viewers can mindlessly consume them in between the actual ad breaks youtube cares about. The more and faster people accept this, the more likely we get improvements.
It took over a year for youtube to agree to let people swear in videos without punishment, including videos that are set as "These are not intended for children", after they unilaterally introduced this swear downranking several years ago.
Youtube cares about Mr Beast and that's about it. If you do not run your channel like Mr Beast, youtube hopes you die painfully and think about your mistake. Youtube actively drives creators making actual art, science, knowledge, and other types of videos to burnout, because Youtube considers creators to be a renewable resource worth exploiting, because there will always be 15 year olds who want to become influencers.
It is not "deeply invasive tracking" or "programmer thinking", it's entirely business. Google's business is ad views, not video views. They want to measure what they care about
And just the fact that if users have a privacy extension blocking the view tracker, is it not just respecting their wishes to not be tracked?
There are defensible business reasons for this, in having a contract already in place at the old CPM, so being unable to double the CPM and half the views mid-contract... but still pretty much fraud.
How is that not substantive?
Some countries have property taxes where you declare the value and the government retains the right to purchase the property for that value for example.
My first thought was to make the advertising cost driven by revenue on the site. But that just reverses the incentive.
In others you'd want, say, auditing or independent third-party verification.
In this case, perhaps an audit involving the deliberate injection of a mix of legitimate and bot traffic to see how much of the bot traffic was accurately detected by the ad platform. Rates on total traffic could be adjusted accordingly.
This of course leads to more complications, including detection of trial interactions, see e.g., the 2010 VW diesel emissions scandal: <https://en.wikipedia.org/wiki/Volkswagen_emissions_scandal>, or current AIs which can successfully identify when they're being tested.
That's a subset of multi-party decisionmaking situations, though it's a useful one to keep in mind.
Altering data would mess with everything. Why is unverified traffic increasing? What's wrong with new marketing efforts? Marketing just requires fixed definitions. e.g. if you have 97% bots but it remains constant that's okay. I know I am spending $x to get $y conversions. I can plan over time, increase or decrease and I can plan. I won't be willing to pay as much as with 0% bots (will pay far far less) but I can set my strategy on this.
It's not that it's x% bots that is the problem. Growth team doesn't adjust strategy on percentage-bot. Growth team adjusts strategy based on return on ad spend. If 0% bots but no return, way worse than 5x ROAS with 99% bots.
Basic ad ops has ad buyers buy ads from different vendors, track which converts (attribution, which has flaws, but generally is decent signal), and allocate spend via return on ad spend. So it hurts the vendor at least as much as the buyer by inflating the cost per action / damaging roas.
Did something similar at a small company I was working at. The VP of marketing sat me down and told me to do the same thing.
After the meeting, I was told by another dev that the VP was tying a monetary value to specific clicks and if I was filtering out the bots, it would make his data look bad and reduce the amount of potential revenue for the company he was touting.
I think you can see how the bots were actually helping him promote how awesome a job he was doing with our web properties to the owners.
I worked for one of the mag7 doing customer support bot tech. Clients internal metrics around containment consistently showed better results than ours - even though you'd normally expect them to be putting pressure on their vendor. because it was a kpi for their internal team to look good to their bosses
It’s not as honest, but more palatable unfortunately
Why not use Microsoft Clarity?
> The bots are getting creepily good at mimicking engagement.
You would think Google Analytics would help a lot with this, but they seem to not care.
Not just not care, but don't they have incentive to report higher traffic?
Answer: pride, curiosity, inspiration, creativity, a sense of ownership, integrity, control, vision
I can't believe these concepts are so foreign to so many.
Like, asking if there's a real technical reason.
But, go off I guess.
The word "revenue" above doesn't sound right. Did you mean "ad impressions", "clicks", "visits", or something else?
If the bot traffic is giving you 40% more revenue, well that's excellent news, right? If it's revenue, then keep doing it.
How does the advertising money get to them to make it worth their while to run bots at such a scale?
I mean sure, I suspect that people in the advertisers are doing it, but surely thats massively risky. there must be a grey market for this kind of transaction?
I agree, I have a hard time understanding the motivation for this behavior. It's obviously happening, I just don't know how someone benefits from a bot pretending to browse the internet.
I was thinking about ad traffic from one of the meta/twitter/tiktok products and couldn't think what the motivation would be
there is also a massive industry of fake accounts and fake engagement for social media and SEO (google). bots are designed to create plausibly real engagement, which is used to trick ranking algorithms into boosting content. these bots have to be real enough to bypass platform detection. clicking through on ads is a way of incentivizing platforms not to shut them down and possibly improving the ranking results, working with the theory that platforms give stronger weight to engagement signals from clients that generate more revenue.
With that said, what do you think the error is here? Could your script have had some false positives? Enough to move the dial?
Repeated purchase attempts, particularly on the cheapest product in the catalogue is a bad sign.
Strange that the item cost $47 and the number of sales was 47.
I was never really a punk-rebel kid, but a certain part of me rooted in the optimism of the early-internet kinda wants to see ad-models crash and burn.
Even advertising "working normally" always had a psychic odor of exploitation and deceit. Ex: "You absolutely need this product or else your peers will hate you."
You could even install them on some digital glasses for a kind of They Live sunglasses situation [0], but please never wear such things while driving...
[0] https://www.youtube.com/watch?v=wqrrh86WDG0
> During my investigation, a source from the e-commerce data industry provided a crucial piece of the puzzle. He explained that his former company was responsible for scraping 70 million retailer web pages every single day. This is a legitimate and massive source of automated traffic.
> Why do they do this? For vital business intelligence. Major retailers like Amazon do not always notify vendors when they run out of stock. So, brands pay for data scraping services to monitor their own products. These "good bots" check inventory levels, see who is winning the "buy box," ensure product descriptions are correct, and track search result rankings. They even scrape from different locations and mobile device profiles to analyze what banner ads are being shown to different audiences.
I think a lot of the players involved in this would say those are bad bots. Having your competitors scrape your site for data would probably be something most website owners wouldn’t like, but getting data about THEIR competitors would be something they WOULD like.
All bot traffic is good from SOMEONE’S perspective, otherwise it wouldn’t be happening. Someone had to program the bot, someone has to be running it. They obviously think they are good bots.
The people running AI scraping bots think they are good bots. Many content creators think those are bad bots. Price comparison sites scraping retailers think their bots are good. The sites being scraped often think they are bad bots.
I just don’t think we can clearly separate good and bad bot traffic without specifying whose perspective we are talking about.
"Bad bots" would be ones that primarily support grift. They siphon money from advertisers, or get other people to waste advertising $, or game product listing metrics so that the bots' owners products get better placement, or so that other manufacturers' products get worse placement. They're not creating any new value; they're removing value from the economy by causing bad allocation of resources.
For example, bots that scrape content for AI training; is that good or bad for a healthy economy? AI can be productive, but is it ‘stealing’ other’s productivity, which could hurt in the long term if it causes decrease in future human production because of diminished rewards?
It's not even a question of whether, in that particular case, AI will displace some forms of human labor, even if that causes a macroeconomic crisis where society needs to figure out a better way to operate in light of new AI-driven economic realities.
It's a simple question of whether internet bots are causing economic inefficiencies and mis-allocation of resources in the classical sense.
The bad bots, discussed in the article, are. AI scraping bots are not. All it takes to see this are all the mentions of fraud in this HN thread. Bad bots are, at their core, fundamentally fraudulent in how they interact with sites. That fraud is compounded into how companies realizing the extent of that fraud react: by hiding or lying to investors and boards, because the fraudulent bot activity has significantly altered market perception. Nothing about pure scrapers is fraudulent, regardless of whether their ultimate goals are good or bad for the economy or humanity, if that could even be known.
This case seems "good" by most perspectives.
Bot traffic has been an issue for years. It has given rise to a new cottage industry of ad fraud detection services, none of which I’ve found particularly valuable. It always comes down to “so what do we do about it?” and no one seems to know how to get bots to stop viewing or clicking on paid media placements. Consumers use Google search and are on FB, IG, TikTok, LinkedIn, etc., and there aren’t really competing ad networks with “fewer bots”. So they keep buying fake traffic knowing that a significant chunk of it is invalid.
I don’t see anything changing unless big tech companies making billions in ad revenue have a large enough incentive to do so. At the moment they have plenty of incentives to keep things as they are.
“Half of the money I spend on advertising is wasted; the trouble is I don’t know which half.” - John Wanamaker
[1] https://jimalytics.com/industry/youre-buying-bots-an-inconve...
But maybe not legislation, because it might upset a ton of people. Any time a social media site goes through and seems to get rid of bots, people complain about how many "followers" they've lost. The fake user/clicker situation is quite pervasive, and a lot of parties, not just the advertising networks, benefit from the inflated numbers.
But it can be so insidious..."wow, that video got 1M YouTube views? Must be very popular!" Orrrr a lot of those views were from bots? Who knows?
So maybe a better approach than legislation is to talk about how the ad/bot fraud can both help us and hurt us and not demonize one side but see how we all may be implicated in it somehow. Maybe that will help us to be more aware of the problem and not fight against people, but try to work together to solve it.
This is not likely to distort the basic numbers in your story, but it makes the premise questionable
If your script can correctly segregate bot traffic from human traffic, and website operators can conclude that bot traffic does not make purchases, then what -- you still don't want to blackhole that traffic right ?
I've nicked that sentence to use as the title above, since it's more neutral and representative of what the article is about.
The numbers that Google Ads shows as invalid clicks is dramatically different from what our platform detects.
In our cases, most fraudulent clicks come from mobile devices and mobile networks. Also, bots don't load page resources in full, making them easier to detect.
If your organization is struggling with fraud click detection, don't hesitate to reach out by email. We are ready to help.
1. https://github.com/TirrenoTechnologies/tirreno
In Google Ads you can just turn off the option to run your ads on non-Google sites; I think it's called their Display Network. Just run your campaign only on Google's search pages.
I'm surprised the article doesn't mention this rather common solution.
I'm none of those things, 10x or otherwise, and would be ashamed if I were. I do run a personal website, though, and most of the traffic is bots that ignore robots.txt. I wish those developers, founders, and marketers would stop doing that to me.
Too often I have been working on ecommerce gigs where the traffic comes from internal 'SEO' tools. Normally the person in charge is marketing, not technical, so it has always been difficult to get past identifying the problem to fixing it.
Often ecommerce companies are very siloed, so one person in one part of the 'team' is scraping the site with their special tools, only for another person to be doing another scrape with their special tools. You can have the guy doing the newsletter doing his thing, the guy doing organic search doing their thing, the guy doing paid ads doing their thing and someone in sales doing their own thing.
The sad thing is that they are typically just a few SQL joins away from exactly the data they want in a format they can digest. However, due to silo-ing, it can be hard to have that conversation.
On top of that, you do get new bots that need to be dealt with. The Huawei bot will scrape everything yet the store might not be delivering to China. So there are legit bots not doing ad fraud that need to be dealt with.
What I also find interesting is that nobody is interested in the server logs. They come for free, and, although not having CDN cache hits, they still record checkout transactions and any pages that don't have a freshly cached page.
Ad fraud also goes on in companies. I worked for a very successful company once and we only measured sales and what was out of stock. We didn't need open rates for emails or click through rates, our main problem was selling too much, which was a nice problem to have. Note that if you sell too much then you aren't going to get it all out the door in a timely fashion, or you run out of big lorries to put the orders in.
Since then I have not worked on a site that is as successful. Instead we have people getting praise for 'false metrics'. Anything an SEO person creates or a marketeer measures will always have some nonsense aspect to it. Or the accounting is mixed with brick and mortar sales even though free shipping and discounts have been given on each sale, with adwords used to get people through the 'door'.
Since sales manager has to report to someone on the board, if the sales numbers aren't good, the nonsense stats can be used to obfuscate the facts. The board only ever care about profit, so I don't like the way this goes down with false metrics of nonsense.
Another fundamental problem in ecommerce is a lack of basic salesmanship. If you work the shop floor doing specialist sales where you have to listen to the customer's needs, then you gain experience in the art of sales. You haven't got to be good at it, in fact it can be better to know your limitations, for me that is big ticket items where I don't have the product knowledge, however, I could always hand those sales over to a much more capable colleague.
You don't win every sale, but, in retail, you can have some really good streaks where no customer leaves empty handed. Your conversion rate is going to be more like 90% in face to face sales if you have the right product at the right price, with customers that don't buy coming back the following day or week to splash the cash.
In High Street retail there is no way you give customers 15%+ off just for stepping through the door. Yet this is table stakes in ecommerce, particularly for small to medium size shops. This instantly devalues the product.
Often there will be chatbots for whatever reason, and I am sure the likes of Dell can get that right, but your typical small ecommerce site will fluff this up too, so any customer daring to use the chatbot will not get instant help from a sales person.
So what to do?
It depends on your product, however, the goal is to get customers for life, not to churn through them. To achieve this it comes down to product, price, availability, shipping times, customer service and incentives for the customer to advertise for you, with reviews, word of mouth and all that hard stuff that needs real human skill. Sometimes it will only be a one-off sale, for example, if someone is buying a mattress. But, even then, the customer service basics matter.
What is also silly is how, with everything you buy, you will get adverts and incentives to buy what you have just purchased. I don't know why this is not considered career limiting, but nobody seems to have fixed this.
Next time, I will find out what data everyone needs and have my own script to collect just that data, then have a live backup that can be used for all internal purposes such as report generation. There will be no mystery script containers downloading 150 scripts with cookies, outside of developer supervision. My hunch is that the only numbers that really matter are sales.
As for why the sales aren't happening, so long as functionality is as it should be on all devices, then you have to dig deeper, to the knowledge level that only someone that has worked the showroom floor understands. Really, a website should be the sales person's knowledge condensed into HTML. Yet nobody asks the guy serving customers what they upsell or what product they recommend if a customer baulks at a given product. Instead we have mystery-meat AI scripts that manage these things.
Thanks for the heads up on ad fraud, normally I am a long way from that due to internal scraping efforts being far more damaging to website performance, and, as a developer, page load times matter to me far more than how much was paid for the traffic.
I guess there is no incentive in most markets. Facebook, etc make only a token effort to reject non-troublesome bot traffic.
[0] bots and other automated traffic which cannot generate revenue or human ad views
Yes, if you only count purchase/sale conversions
Maybe no, if you also count clickthrough and view conversions, perhaps even lead conversions sometimes (because fake sign ups are possible).
But you’re right. Purchase conversions are one incentive
I'd love to have an agent that goes online whenever I'm running low on toilet paper or something, browses all the stores and clicks all the ads, and automatically orders the best deal it can find.
The irony of complaining about bot traffic with bot copy is too funny.
> It was not a sophisticated piece of software, just a tool designed to observe how "users" actually interacted with the website. I was not just counting clicks; I was watching behavior
I'm working right now on an inventory management system for a clinic which would really benefit from pulling the prices and availability from a very specialised online shop. I wish I could just get a large, fully cached status of all items in a json/CSV/whatever format. But they're extremely not interested, so I'm scraping the html from 50 separate categories instead. They'll get a few daily bot hits and neither of us will be happy about it.
If people are scraping data that you're not selling, they're not going to stop - just make it trivially accessible instead in a way that doesn't waste resources and destroy metrics.
Sure you might be willing to build the bot to scrape it... but some other competitors won't go to this effort so it still means a bit of information asymmetry and stops some of your competitors poaching customers / employing various marketing tactics to exploit short term shortages or pricing charges etc.
> 'Why hand your competitors data on a silver plate'?
To lessen the issue from the article and free up server resources for actual customers.
A simple and reliable API? That's different.
Bot prevention measures can be good, but the more hoops you make your users jump though (CAPTCHA etc), the more legitimate users will drop off. Those have significant impacts on conversion rates.
I would think fixing this should involve the analytics and attribution side rather than adding friction to your e commerce flow.
Especially as bot tech continues to get better and more indistinguishable from real traffic.
The second is best "blocked" by doing nothing at all and instead silently dropping them from analytics.
That said, 73% doesn't come as a surprise. If anything I expect it to be higher.
I guess this quote sums up the situation
> When I tried to bring this up with a few major ad platforms, the conversation always followed a predictable script. The sales reps were incredibly friendly until I mentioned click fraud or bot traffic. Then, the tone shifted instantly to corporate-speak: "Our AI detection is industry leading" and "We take ad fraud very seriously." It was a polite but firm wall, a clear signal to stop asking questions.
> One rep I had known for years finally admitted the truth off the record. "Dude, we know," he said. "Everyone knows. But if we filtered it all out properly, our revenue would drop 40% overnight, and investors would have a meltdown."
Lawsuit time?
If the SEC weren't weak maybe they could go after them for fraudulent numbers.
It just doesn’t ring true. That aspect of the story isn’t novel at all, and someone in that line of work should surely have known all this, right?
Now the section on categorising different bot patterns, that’s more interesting, and I haven’t seen so much said about it.
For most consumers, it’s entertainment, but for industrial use it’s great for this kind of fraud. And very difficult to detect at scale, since the only cost effective tools for this kind of analysis are also ML/AI, and hence can be fooled more predictably/trained against.
144 more comments available on Hacker News