Building a Trust & Safety Program from Scratch: Lessons from Amazon, Google, and TikTok

Every successful platform eventually needs Trust & Safety. The question is whether you build it proactively or reactively after a crisis.

I've built Trust & Safety operations from scratch twice—at TikTok LATAM (zero to regional safety infrastructure) and within new product launches at Amazon and Google. Here's what actually works when you're starting from nothing.

The Biggest Mistake: Waiting Too Long

The most common pattern I see:

  1. Launch: "We'll deal with safety later"
  2. Early growth: "We're too small to have real safety issues"
  3. Inflection point: First serious incident, regulatory inquiry, or press coverage
  4. Panic: Scramble to build safety infrastructure under crisis pressure
  5. Over-correct: Spend 2-3x more than if you'd built it proactively

Reality: It's much cheaper to build safety infrastructure before you need it than to retrofit it during a crisis.

When to Start Building Trust & Safety

Bare minimum triggers:

Ideal timing:

Too late:

Phase 1: Foundation (Weeks 1-4)

Start with the minimum viable safety infrastructure:

1. Define What "Unsafe" Means for Your Platform

Write down specific prohibited content and behavior for your platform.

Start with legal minimums:

Add platform-specific harms:

Document as policy: Create community guidelines or acceptable use policy

Example (Dating App):

2. Build User Reporting Mechanism

Users need a way to report safety issues.

Minimum viable:

Tools to use:

Don't build custom tooling yet. Use existing systems.

3. Establish Review Process

Someone needs to review reports and take action.

Early stage (<10,000 users):

Response time target:

Create decision log:

4. Set Up Basic Detection

Even pre-launch, implement basic automated detection:

Technical controls:

For images/video:

For text:

Realistic early stage: 90% reactive (user reports) + 10% proactive (automated detection of extreme content)

Phase 2: Scaling Infrastructure (Months 2-6)

As you grow from thousands to tens of thousands of users:

1. Formalize Moderation Operations

Hire your first Trust & Safety person:

Typical first hire profile:

Salary range: $80K - $120K depending on market and experience

2. Implement Moderation Vendor (If Needed)

When to use moderation vendors:

Major vendors:

Typical pricing: $8-$15 per hour per moderator (offshore), $25-$40 (US-based)

Hybrid model (recommended):

3. Build Automated Detection (v2)

Invest in better proactive detection:

Text classification:

Image/video detection:

Behavioral signals:

Don't over-automate: Keep humans in the loop for final decisions on removal

4. Create Appeals Process

Users will disagree with your moderation decisions. You need a way to handle appeals.

Basic appeals process:

  1. User clicks "Appeal" on enforcement notification
  2. Form asks: why do you think this was a mistake?
  3. Different reviewer (not original moderator) re-reviews
  4. Decision within 48-72 hours
  5. Overturn if wrong, uphold with explanation if correct

Track appeal metrics:

5. Establish Incident Response Plan

You will have safety incidents. Plan before they happen.

Define severity levels:

For each severity:

Practice: Run incident simulations quarterly

Phase 3: Mature Operations (Months 6-18)

As you scale to hundreds of thousands or millions of users:

1. Build Cross-Functional Safety Processes

Trust & Safety can't be siloed. Integrate with:

Product: Safety review for all new features

Engineering: Safety infrastructure roadmap

Legal: Regulatory compliance alignment

Communications/PR: Crisis communication

2. Implement Quality Assurance

How do you know your moderation is accurate?

Sampling and audits:

Calibration sessions:

User feedback:

Target accuracy: 92-96% (perfect accuracy impossible with subjective policies)

3. Develop Transparency Reporting

Even before it's legally required (DSA, etc.), build transparency reporting:

Track:

Publish: Quarterly or semi-annually

Benefits:

Common Mistakes When Building Trust & Safety

Mistake 1: "We're Different, Normal Rules Don't Apply"

I hear this constantly: "We're a professional network, we don't have safety issues" or "Our users are vetted, we don't need moderation."

Reality: Every platform with humans has safety issues. LinkedIn has fraud and harassment. GitHub has harassment and spam. Professional platforms aren't exempt.

Solution: Assume you'll have safety issues and build accordingly.

Mistake 2: Over-Reliance on Automation

"We'll just use AI to moderate everything automatically."

Reality: Current AI can help detect, but shouldn't auto-remove except in narrow cases (exact CSAM hashes, known malware). Humans needed for context and edge cases.

Solution: Automation flags for human review. Humans make final decisions.

Mistake 3: Vague Policies

"Be respectful" and "don't abuse the platform" are not enforceable policies.

Reality: Moderators and users need specific definitions. What exactly is harassment? What's hate speech? What's spam?

Solution: Write specific, enforceable policies with examples. "Harassment includes repeated unwanted contact after a user blocks you" is enforceable.

Mistake 4: No Policy Enforcement for "Important" Users

"This user drives a lot of engagement, we can't ban them even though they violate policies."

Reality: Inconsistent enforcement destroys trust and creates legal risk. If your policy says X is prohibited, you must enforce on everyone.

Solution: Enforce policies consistently. If an important user is valuable despite policy violations, change the policy—don't make exceptions.

Mistake 5: Treating Safety as Cost Center

"Trust & Safety doesn't make money, minimize spending on it."

Reality: Safety failures cost more than safety infrastructure. One major incident can cost millions in legal fees, settlements, lost users, PR damage.

Solution: Treat safety as risk mitigation. The ROI is avoiding catastrophic losses.

Need Help Building Your Trust & Safety Program?

Echelon Advisory helps startups build Trust & Safety infrastructure from scratch, based on lessons from launching operations at Amazon, Google, and TikTok.

Services:

Contact Us

Key Takeaways

Building Trust & Safety infrastructure proactively is one of the best investments a startup can make. It's cheaper, faster, and less stressful than building it reactively during a crisis.


About the Author

Maneesha Pandey is the founder of Echelon Advisory Services, specializing in Trust & Safety, AI Governance, and EU regulatory compliance. She spent 14+ years building Trust & Safety infrastructure at Amazon, Google, and TikTok, including launching TikTok's LATAM Trust & Safety operations from scratch.

Learn more about Echelon Advisory Services