Twilio Protocols

Enforce good data, always.

Take control of your data quality with intuitive data controls and real-time validation, built for enterprise scale.

Flowchart showing a tracking plan connecting Android, Node.js, iOS to messaging, warehouse, advertising, sales/support.

How Twilio Protocols works

Diagram showing eCommerce mobile tracking plan with user data connecting to various departments.

Protocols helps businesses automate several workflows and processes to maintain clean and consistent customer data, at scale.

  • Create a shared data dictionary

    Replace outdated spreadsheets and align teams throughout your company with a well-designed Tracking Plan.

  • Diagnose quality issues

    Automate the QA process and always catch data quality violations—without wasting time manually testing your tracking code.

  • Maintain trust in your data

    Enforce data standards with strict controls to block non-conforming events and ensure high-quality data over time, so you can be confident in your data.

Protocols use cases

Protect data integrity with data quality best practices.

Develop an actionable Tracking Plan

Establish a central source of truth for product, engineering, analytics, and business teams. Define objectives, formalize naming and collection standards, identify users, and set tracking events and properties.

Creative Market used Protocols to develop a Tracking Plan to validate customer data before it’s distributed downstream. They’ve reduced time spent resolving data issues, saving their engineering resources for core product development.

Laptop computer with security lock
93%

faster detection of data issues

Stop watch
24

hours or less detection time

Protocols features

Access tools to scale your data quality management.

Profile of Jenna Turner showing her email and various professional data integration tools.

Connect all your customer data

Maintaining clean and consistent data is easy when you collect, unify, and activate all your first-party data in a single platform.

  • Tracking Plan

    Replace outdated spreadsheets with a data spec that outlines the events and properties you intend to collect from your sources.

  • Centralized violation dashboard

    Review, resolve, and forward non-conforming events from all connected sources in one place.

  • Customized schema controls

    Configure schema to selectively block events, properties, and traits at the source, or prevent them from passing through your CDP.

  • Typewriter API

    Generate strongly-typed analytics libraries based on your Tracking Plan and reduce implementation errors.

  • Transformations to fix bad data

    Apply transformations for total control over what data gets sent to downstream destinations.

  • In-app reporting and daily email digests

    Get the information you need to quickly take action on every invalid event before your team accidentally uses bad data to make decisions.

  • Anomaly detection

    Receive notifications when event volume anomalies or violation counts occur to know when your data collection is broken, missing, or incorrect.

  • Segment’s Public API

    Extend your workflow around customer data collection and activation with our Public API.

Start automating and scaling data governance

# Download the helper library from https://www.twilio.com/docs/python/install
import os
from twilio.rest import Client


# Find your Account SID and Auth Token at twilio.com/console
# and set the environment variables. See http://twil.io/secure
account_sid = os.environ['TWILIO_ACCOUNT_SID']
auth_token = os.environ['TWILIO_AUTH_TOKEN']
client = Client(account_sid, auth_token)

verification = client.verify \
                     .v2 \
                     .services('VAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \
                     .verifications \
                     .create(to='+15017122661', channel='sms')

print(verification.status)

Need help setting up Protocols?

Work with one of our trusted partners to set up your data governance solution and start automating data quality. View partners

Taking control of your data quality is easy

Discover how Protocols can improve trust in your data, reduce time spent validating, and ultimately allow your business to grow faster.

Woman in an orange sweater smiling while looking at a laptop screen.

FAQ

Data cleansing is the process of ensuring data is complete, accurate, and reliable so that it can be used for analysis, decision-making, and reporting. Data cleansing includes correcting outdated or missing information, applying standardized naming conventions to data entries for consistency, deduplicating events, recording data transformations for transparency, and more.

Without properly cleaning data, organizations run the risk of basing strategies and campaigns off inaccurate or misleading information. Duplicate data entries, incomplete fields, and inconsistent naming conventions are common culprits behind “bad data,” which can cost businesses millions of dollars each year in misguided insights and poor decision-making.

To build a strong data foundation, you need to pay close attention to the way you prioritize what data you’re going to be ingesting, the naming conventions for that data, and proactively aligning data collection with privacy. The first step to this is identifying outcome oriented data based on the business goals your organization wants to achieve with a CDP. You absolutely need to define these goals first if you want to track meaningful data with your CDP.  

Once you’ve identified outcome-oriented events to measure, you can take that and build a Tracking Plan. A Tracking Plan gives you a consistent roadmap for what data is important, where it’s being tracked, and why you’re tracking it. For an in-depth breakdown on aligning people and processes around a tracking plan, we recommend reading the Customer Data Maturity Guide.