In the vibrant cultural landscape of Quebec, connecting audiences with events, exhibitions, and performances isn't just about listing what's available—it's about understanding how people engage with these opportunities. When our team set out to build a centralized cultural platform, we knew we needed more than just a catalog; we needed insights into how our partner organizations' content was performing.
This article walks through how we built a lightweight, high-performance URL tracking system that gives us valuable analytics while maintaining lightning-fast performance for users.
The Challenge: Invisible Tracking Without Slowing Things Down
Our cultural platform serves as a hub for hundreds of partners across Quebec—museums, theaters, festivals, and more. Each partner displays content on their own websites using our data, including images and ticket links. We wanted to understand how users interact with this content, but faced several challenges:
- Performance requirements: Any tracking solution needed to be virtually invisible to end users—no noticeable delays when loading images or clicking links
- Scale considerations: During major festivals, our system might handle millions of redirects in a short timeframe
- Detailed analytics: We needed to track not just clicks but which partners were generating engagement
- Technical simplicity: The solution had to be easy for partners to implement
The Solution: Edge Computing + Lightweight Redirects
After exploring several options, we landed on a elegant solution: A URL redirection service built on Cloudflare Workers. Here's how it works:
The User Flow
- Partner websites use special URLs for our content:
- When a user's browser requests an image or clicks a ticket link, our Cloudflare Worker intercepts the request
- The Worker logs the interaction (an impression for images, a click for tickets)
- The Worker redirects the user to the actual resource almost instantly
- The entire process happens in milliseconds, with no perceptible delay
https://lavitrine.com/r/partnerId/img/imageId
https://lavitrine.com/r/partnerId/ticket/ticketId
Why This Approach Works Well
This approach gives us several key advantages:
Edge Computing Speed: Cloudflare Workers run on servers physically close to users, providing response times often under 5ms
Minimal Overhead: The redirection adds negligible latency (typically 50-100ms) while providing valuable tracking data
Partner Simplicity: Partners only need to use our URL format—no JavaScript to install, no cookies to manage
Scalability: The system scales automatically during high-traffic periods like summer festivals or holiday seasons
Under the Hood: Technical Implementation
For the technically curious, here's how we built the system:
The Architecture
Our tracking system consists of a few core components:
- Cloudflare Worker: A small JavaScript function that handles URL parsing, logging, and redirection
- KV Storage: A distributed key-value store that holds both the URL mappings and temporary analytics data
- Analytics Pipeline: A background process that aggregates and processes the tracking data
- PostgreSQL Database: The final destination for our analytics data, optimized for time-series analysis
The Code
sequenceDiagram
autonumber
participant User
participant Partner as Partner Website
participant Edge as Cloudflare Worker
participant KV as KV Storage
participant Content
participant Analytics as Analytics Pipeline
User->>Partner: Visits cultural page
Partner->>Edge: Request content
/r/partnerId/img/123
Edge->>KV: Look up target URL
KV-->>Edge: Return destination URL
Note over Edge: Process request
(~5ms)
Edge->>KV: Log impression/click
with partner attribution
Edge-->>Content: 302 Redirect
Content-->>Partner: Deliver actual content
Partner-->>User: Display content
Note over Edge,Analytics: Asynchronous Processing
KV->>Analytics: Batch process events
(every few minutes)
Analytics->>Analytics: Aggregate by partner,
content type, region
Note over User,Analytics: Total user-facing time: ~50ms
Optimizing for Performance
To maintain the sub-100ms response times required for a seamless user experience, we implemented several optimizations:
- Batched Analytics: Instead of writing each event to the database immediately, we aggregate them in memory and write in batches
- Tiered Caching: URL mappings are cached at the edge for extremely fast lookups
- Adaptive TTLs: Images are cached longer than ticket links, which may contain dynamic availability information
- Fallback Behavior: If analytics logging fails for any reason, we still complete the redirect to ensure user experience isn't impacted
Analytics Insights: What We've Learned
With this system in place, we've gained valuable insights that were previously invisible:
- Partner Performance: We can now see which partners generate the most engagement with cultural content
- Image vs. Ticket Performance: We track not just impressions but conversion to ticket clicks
- Regional Patterns: The system captures anonymized geographic data, revealing regional interest patterns across Quebec
- Festival Impact: During major events like Montreal Jazz Festival, we can track the spike in related content views across all partner sites
Implementation Challenges and Solutions
Building the system wasn't without challenges:
Challenge 1: Handling Peak Traffic
During major cultural events, our traffic can increase 10x overnight. Our solution: Edge computing with Cloudflare Workers automatically scales to handle traffic spikes without manual intervention.
Challenge 2: Data Privacy Compliance
As a Quebec-based platform, we needed to ensure compliance with provincial privacy regulations. Our solution: We anonymize all tracking data and maintain clear data retention policies.
Challenge 3: Partner Adoption
Getting partners to implement new URL structures initially faced resistance. Our solution: We created simple documentation and demonstrated the analytics benefits, leading to enthusiastic adoption.
Lessons for Other Developers
If you're building a similar system, here are key takeaways from our experience:
- Embrace edge computing for performance-critical components like redirects and tracking
- Keep the partner implementation simple to ensure widespread adoption
- Design for failure by ensuring critical user flows (like redirections) complete even if analytics logging fails
- Batch process analytics data instead of making database writes on the critical path
- Use appropriate caching strategies for different content types (static vs. dynamic)
Results: Fast Tracking Without Compromise
The results of our implementation speak for themselves:
- Median redirect time: 47ms
- Analytics coverage: 99.7% of all interactions tracked successfully
- Partner adoption: 93% of partners implemented the tracking URLs
- System reliability: 99.99% uptime since launch
Conclusion: Making Technology Invisible
The best technical solutions often become invisible—working so reliably that users and partners forget they exist. Our URL tracking system achieves this balance: providing valuable analytics insights without compromising the user experience.
For Quebec's cultural community, this means better understanding of how audiences engage with events and exhibitions across dozens of partner platforms, all without slowing down the experience for culture enthusiasts.
The next time you click on a festival image or ticket link from one of our partner sites, know that your interaction is being counted—all in under 100 milliseconds.