Most B2B marketing content is about adding — new channels, new tactics, new frameworks. This Field Note is about subtraction. Stopping the right things frees more budget, attention, and pipeline capacity than starting the next new thing. Here is the logic and the method.
Field Note 009 · Strategy
What should we stop doing?Adding is celebrated. Stopping requires admitting that a past decision was wrong or that the world changed and nobody updated the activity to match. That asymmetry explains why most marketing programmes accumulate inertia activities faster than they shed them.
Starting something new requires energy and conviction. Stopping something requires admitting that a past decision was wrong — or that the world changed and nobody updated the activity to match. Both are politically and emotionally harder than launching the next campaign.
Every activity you keep running has a cost beyond its direct budget: team attention, cognitive load, coordination overhead, and the opportunity cost of not doing something better instead. These costs are invisible in most marketing planning because nobody accounts for the attention tax of maintaining existing activities.
A channel gets added during a growth phase. A report gets created for a one-time leadership request. A campaign runs because it ran last quarter. An event gets attended because "we've always gone." None of these get questioned because stopping requires a conversation that nobody wants to initiate.
In Indian B2B company cultures, questioning a past decision can feel like questioning the person who made it — especially if that person is still in the room. The stopping conversation is harder here. Which makes it more necessary and more valuable when it happens.
L1 questions ask how to prioritise. L2 questions ask what is costing us more to continue than to stop.
Before the framework, a reference catalogue. These are the activities that most commonly persist past their useful life in Indian B2B marketing — and the specific signals that tell you it is time to stop each one.
If your primary report to leadership shows MQL volume, you are optimising for the wrong outcome. MQLs measure marketing activity, not marketing impact. High MQL volume with low SQL conversion means you are generating noise, not pipeline.
Publishing 12 blog posts per month to "maintain cadence" when four of them are genuinely useful is producing eight pieces that dilute your authority and consume production capacity. In B2B, four excellent pieces outperform twelve average ones — in search, in buyer trust, and in sales enablement value.
Every Indian B2B company has two or three events on the annual calendar that "we always attend." If you cannot name three specific qualified prospects you expect to meet there, the event is not a pipeline investment — it is a relationship maintenance activity, and that is a much lower-value use of event budget.
The weekly marketing dashboard with 24 metrics. The monthly channel performance report that gets filed and forgotten. The quarterly brand tracker that nobody changed a decision based on. These reports consume significant production time and create the illusion of measurement without producing decisions.
A Twitter/X account with 800 followers updated twice a week. An Instagram account for a B2B industrial company. A YouTube channel with six videos that have 200 views each. These channels consume content production capacity and produce no pipeline signal — but they persist because "we should have a presence."
Campaigns that generate impressions, clicks, and engagement with no pathway to pipeline are awareness spend masquerading as demand gen. If the campaign cannot produce a qualified conversation within a defined timeframe, it belongs in the brand investment budget — not the demand gen budget. The confusion between the two is where budget gets wasted.
Indian industry awards often consume significant marketing team time in nomination preparation and have low recognition among global buyers. The test is simple: would a US, UK, or European procurement team in your ICP recognise this award and find it credible? If no, it is internal credibility signal, not buyer-facing proof.
A 200-asset content library that sales accesses 12 pieces from is not a content library — it is a storage problem. The 188 unused pieces consume maintenance time, confuse new sales hires, and create the false impression of sales enablement coverage. The pieces with zero usage in 12 months should be removed, not archived.
These six steps produce an explicit, agreed, communicated stop list — not a vague deprioritisation that reverses itself within a quarter. Each step is designed to address a specific reason why inertia activities persist.
Run the activity audit — list everything, including what nobody questions
The first step is complete visibility. Most marketing teams know their planned activities but not their full activity load — including recurring tasks, maintenance work, and inherited programmes that run without appearing in any plan.
Most marketing teams discover they are running 30-40% more activities than they planned for — in maintenance, inherited programmes, and recurring tasks that compound over time. That 30-40% is the attention budget that could be doing something with higher return.
Apply the zero-based test to every item on the list
The zero-based test is simple: if we were starting from scratch today, with the same budget and team, would we choose to run this activity? Not 'should we continue it' — that question has inertia built into it. 'Would we start it today' removes the sunk cost bias.
Zero-based activity review forces the question of whether past decisions were correct. In organisations where the person who made those decisions is still in the room, this is politically sensitive. The way through is to frame every stop decision as a reallocation, not a criticism: 'we are stopping X to invest more in Y, which is producing better pipeline.'
Indian SaaS companies targeting global buyers often inherit a content production cadence from an agency or a previous marketing head that prioritises volume. The zero-based test almost always reveals that four to six deeply researched, buyer-specific pieces per month produce more pipeline than twelve to sixteen generic ones — and require less total production time because the research is shared. Stop the volume. Protect the depth.
IT services firms often maintain a calendar of eight to ten industry events annually — Nasscom, CII, multiple vertical summits, a few international conferences. When run through the zero-based test, most teams find that two or three events produce the majority of qualified conversations. The rest are attended for visibility or relationship maintenance. The visibility value is real but it is not pipeline value, and it should not come from the same budget line. Stop the obligation events. Double the investment in the events that consistently produce qualified conversations.
Many Indian manufacturers run LinkedIn and Google campaigns targeting broad "manufacturing decision maker" audiences that never produce qualified RFQs. The procurement engineers and plant managers who actually make sourcing decisions are not discovering new suppliers through LinkedIn feed ads. They are using trade directories, trade show directories, and specific technical search queries. The zero-based test for digital manufacturing campaigns almost always produces a stop recommendation — the budget produces impressions but not RFQs from qualified buyers.
Pharma marketing teams often list fifteen to twenty KOL relationships but have meaningful engagement with fewer than five. A KOL programme with shallow relationships produces no credibility signal visible to global buyers. The zero-based test: if you were starting a KOL programme today, would you invest in fifteen shallow relationships or five deep ones? Almost always five deep ones. Stop maintaining the relationship fiction with KOLs who have had no meaningful engagement in 12 months. Invest the freed time in the three to five who are genuinely engaged and willing to be public advocates.
BFSI, logistics, and professional services firms often report marketing performance through website traffic, social media followers, and press mentions — metrics that are visible but disconnected from pipeline. The zero-based test for reporting: if you were designing a marketing dashboard today, knowing what you now know about your buying cycle, would you put website visitors on it? Almost never. Stop reporting metrics that don't connect to pipeline decisions. Replace with pipeline quality metrics that finance and sales can act on.
Identify the specific inertia activities — and name why they persist
Inertia activities have a signature: they have been running for a long time, nobody can point to a specific outcome they produced recently, but they continue because stopping them requires a conversation nobody wants to have. Identify them and name the real reason they persist.
| Activity type | Inertia signal | Evidence needed to keep | Default if no evidence |
|---|---|---|---|
| Content channel (blog, newsletter, social) | Running for 12+ months with no pipeline attribution | At least one qualified inbound per month traceable to this channel | Stop or pause |
| Industry event | Attended annually with no post-event pipeline review | Named prospects met + pipeline created within 90 days for two consecutive years | Remove from calendar |
| Marketing report | Produced regularly but decisions don't reference it | One documented decision that changed because of this report in the last quarter | Stop producing |
| Paid campaign | Running with no SQL attribution for 90+ days | At least 2 SQLs per month with positive CAC payback | Pause and reassess |
| Sales enablement asset | In library but not used by sales in 12 months | Evidence of usage in active deals in the last quarter | Remove from library |
| Award or accreditation | Pursued annually with no buyer-visible impact | Named buyer who cited it as a shortlist factor | Stop until evidence exists |
| Agency retainer | Active but outputs not connected to pipeline | Pipeline metric that changed as a result of agency output | Renegotiate or cancel |
| Tool subscription | Paying monthly but team rarely uses it | Measurable outcome that changed because of this tool | Cancel |
Inertia activities that persist for political reasons should be named as such. 'We continue this because the CEO started it and stopping it feels like a criticism' is a real reason — and naming it honestly is the first step to addressing it. Vague avoidance of the conversation keeps the activity running indefinitely.
Calculate the real cost of continuing each inertia activity
The direct budget of an inertia activity is only part of its cost. The full cost includes team attention, coordination overhead, and — most importantly — the opportunity cost of what those resources could be doing instead. Calculate the full cost before making a stop decision.
When the cost of continuing an inertia activity is expressed in rupees and hours — not in vague terms like 'it takes some time' — the stop decision becomes easier to make and easier to defend. A report that takes 12 hours per month to produce and has not influenced a single decision in six months is costing ₹24,000 per month in attention. That number changes the conversation.
Build an explicit stop list — written, agreed, communicated
Stopping is not the same as deprioritising. Deprioritised activities return. Activities on a written, agreed stop list with named owners and a communication plan actually stop. The stop list is a formal document, not a vague intention.
Activities that are 'deprioritised' rather than explicitly stopped tend to resurface within one or two quarters — because the team still feels the obligation, because a stakeholder asks about them, or because inertia reasserts itself. The explicit stop list removes ambiguity: this activity is over, these are the resources freed, this is what they are going to instead.
Manage the stopping — internally and externally
Stopping activities produces reactions. Stakeholders who expected a report won't get it. Team members who owned an activity need a new direction. External partners who depended on a programme need to be informed. Manage these consequences deliberately or inertia activities return within a quarter.
SaaS companies with growth-oriented cultures can find stopping harder than starting because stopping looks like pulling back. Frame every stop decision as a focus decision: "We are stopping these three activities to go deeper on the two that are producing qualified pipeline." The reallocation story — what the freed resources will do — is as important as the stop decision itself. Show the pipeline logic, not just the cost reduction.
IT services firms with long histories in certain events or industry associations face a specific stopping challenge: some activities are relationship maintenance for partnerships or industry standing that has non-pipeline value. When stopping these, distinguish clearly between "we are stopping because it produces no pipeline" and "we are reclassifying this as a relationship investment with a different budget line and different success criteria." The former is a stop. The latter is a reclassification. Both reduce the demand gen budget confusion.
Manufacturing companies that stop co-marketing programmes with distributors or channel partners need to manage those relationships carefully. The stop conversation with a distribution partner is different from an internal stop decision — it requires advance notice, a clear explanation, and where possible a replacement programme that serves the partner's interests differently. Stopping a programme without this communication can damage a distribution relationship that produces more pipeline than any direct marketing activity.
Pharma marketing activities that have medical affairs involvement require medical affairs to be part of the stop decision, not just the start decision. Stopping a KOL engagement programme or a scientific conference sponsorship without medical affairs awareness can damage relationships or create regulatory obligations. Include medical affairs in the stop list review for any activity with a scientific or clinical component.
In BFSI, logistics, and professional services, some reporting activities exist to satisfy leadership curiosity about brand metrics — press mentions, social followers, website rankings — that don't connect to pipeline. Stopping these reports requires a conversation with leadership about what marketing is being measured on. That conversation is harder than stopping the report itself, but it is the necessary step: if leadership expects vanity metric reporting, the team will produce it regardless of what the stop list says.
Stopped activities re-emerge when the stop is not communicated clearly enough, when a new team member inherits a list of 'what we do' that includes it, or when a stakeholder asks about it and the team restarts it to avoid the conversation. The 90-day review and a permanently updated activity list prevent this.
How companies across SaaS, IT services, manufacturing, and other B2B categories made stop decisions — and what changed when the inertia activities were removed.
HubSpot famously ran an experiment that most content teams would find counterintuitive: they significantly reduced their blog publishing frequency and invested the freed resources in updating and improving existing high-performing content. The result was an increase in organic traffic and qualified inbound. The insight was that a large volume of average content competed with their own high-quality content for search visibility and reader attention. Stopping the average content to protect and deepen the best content was a subtraction decision that produced better outcomes than continuing to add. For B2B companies running high-volume content programmes, the stop decision is often the highest-leverage marketing decision available.
A mid-sized Indian IT services firm attending nine industry events per year ran each event through the zero-based test and pipeline review process. The analysis showed that two events — Nasscom and one vertical-specific conference — produced 80% of their event-sourced pipeline. The remaining seven were maintained for "visibility and relationship reasons" that could not be connected to a specific business outcome. Four were stopped outright. Three were reclassified as relationship maintenance activities with a separate budget and no pipeline expectation. The freed budget of approximately ₹25L per year was redirected to an account-based marketing programme targeting 20 named enterprise accounts. The ABM programme produced three times more qualified pipeline in its first year than the seven stopped events had produced collectively in the previous two years.
A B2B SaaS company was producing a 24-metric weekly marketing dashboard that took six hours per week to compile. An audit revealed that the board and leadership team referenced three of those metrics in actual decisions — pipeline created, SQL conversion rate, and CAC by channel. The remaining 21 metrics were produced because they had always been produced. The dashboard was replaced with a five-metric pipeline quality report: pipeline generated, SQL rate by channel, win rate by source, CAC payback, and expansion rate by acquisition channel. The report took 90 minutes per week to produce. Board marketing conversations shifted from "what do these numbers mean" to "where should we invest more." The stopping decision freed four-and-a-half hours per week and improved the quality of every marketing conversation with leadership.
An Indian precision parts manufacturer had been running LinkedIn brand awareness campaigns for eight months targeting "procurement managers in automotive and industrial manufacturing." The campaigns produced 3,400 impressions and 40 page clicks per month. They produced zero attributed RFQs. A zero-based test and buyer interview programme revealed that the company's target buyers — procurement engineers at European and US automotive OEMs — discovered new Indian suppliers primarily through Thomasnet listings, industry directory searches, and trade show catalogues, not LinkedIn feeds. The campaign was stopped. The budget was redirected to a comprehensive Thomasnet listing, a trade directory audit, and upgraded trade show booth presence at IMTEX. Within six months, three inbound RFQs from global buyers had been received through Thomasnet alone.
Intercom's marketing team found through CRM document tracking that sales was regularly using 12 out of 180 content assets — a 7% usage rate. A full audit of the library revealed that the 168 unused assets fell into three categories: outdated content that no longer reflected the product, aspirational content that addressed use cases sales rarely encountered, and duplicates that had been created over multiple agency relationships without curation. The 140 low-usage assets were removed. The remaining 40 were reviewed, updated, and reorganised by deal stage and objection type. Sales content usage quadrupled in the first quarter after the cull — not because the new assets were dramatically better, but because the library was navigable and the irrelevant noise had been removed. Less content, used more, produced more sales impact than more content used rarely.
List every recurring marketing activity your team runs — including the ones nobody questions. For each one, write one sentence: what specific outcome has this produced in the last 12 months that would not have happened without it? The activities you cannot write that sentence for are your candidates to stop. You probably have more of them than you expect.
Then calculate the total team hours per week consumed by activities that failed that test. Multiply by your average hourly team cost. That number — the attention tax of your inertia activities — is the real size of your stopping opportunity.
Every month, one hard B2B marketing problem.
First principles thinking. Real India context.