‹ Back
Data Engineer - Privacy Team
JOB SUMMARY
Roles
Data Engineer
Skills & Technologies
Languages:PythonSQLGo
Cloud/DevOps:AWSGCPDocker
Tools:Git
Job details
About IPinfoIPinfo is a leading provider of IP address data, including geo
location, VPN and residential proxy detection, mobile carrier data, and over 20 other context tags.
Our API handles over 120 billion requests per month, and we also license our data for use in many products and services you’ve likely interacted with.
We’re a fast-growing, bootstrapped company with a globally distributed team of around 60 people.
Our data powers customers such as Cloudflare, T-Mobile, SpaceX, DemandBase, and Clearbit, among many others.
We also sponsor and contribute to academic conferences like ACM IMC and ACM CoNEXT, staying closely connected with the internet measurement and research community. About the Privacy TeamThe Privacy team delivers accurate, verifiable insight into how IP addresses are used across the internet.
Instead of relying on opaque risk scores or ambiguous classifications, we focus on defensible, fact-based signals derived from observable network behavior.
Our work centers on identifying and analyzing the applications, protocols, and infrastructure used to anonymize or tunnel internet traffic—such as VPNs and residential proxy networks.
We combine large-scale internet measurement with careful analysis to understand how these technologies are deployed and how they behave in real-world environments. Examples of our research and methodology include:VPN
Location Mismatch Report – examining how VPN exit
locations differ from expected geographies, https://ipinfo. io/blog/vpn-
location-mismatch-reportResidential Proxy Coverage – analyzing the growth and characteristics of residential proxy networks, https://ipinfo. io/blog/residential-proxy-coverage-growsWe collaborate with internet infrastructure providers, including major CDNs and network operators / ISPs, as well as customers across sectors such as banking, ad tech, fraud prevention, and security.
Our work helps these partners understand traffic quality, abuse patterns, and privacy tooling with clarity and confidence. Team members are expected not only to build and maintain systems, but also to challenge assumptions, validate signals, and publish findings that can withstand external scrutiny.
How We WorkWe’re an ambitious, fully remote team spread across the globe.
We sync up on a monthly all-hands Zoom call, and most teams meet every 1–2 weeks. Everything else happens asynchronously using Slack, GitHub, Linear, and Notion.
This setup allows you to choose the hours that work best for you, while still collaborating closely with teammates. Autonomy, ownership, and clear communication are essential to how we operate. ResponsibilitiesDesign, build, and operate data collection and analysis pipelines to detect proxy usage, VPN applications, and related spoofing or evasion behaviorWork with large-scale internet measurement data (we collect 75+ TB per week, including BGP, DNS, ping, and traceroute data from 1200+ global vantage points)Research, apply, and implement techniques from cutting-edge internet measurement and network security researchMaintain a high bar for signal quality and defensibility, prioritizing observable network behavior over heuristics or guessworkCommunicate findings clearly by contributing to blog posts, technical documentation, and research publications, both internally and externallySkills and ExperienceRequiredBackground in one or more of:Internet measurements or network telemetryData engineering or experience with batch, streaming, or real-time data pipelinesNetwork engineering or ISP / CDN operations with experience analyzing traffic behaviorSecurity research or applied network securityThreat intelligence or abuse / fraud analysisDeep understanding of networking protocols and networked applications (e. g. VPNs, proxies, tunneling, routing behavior)Proficiency in Bash scripting, Python, Go, or similar languages for building data collection and analysis systemsProficiency in SQL for querying, analyzing, and validating large datasetsProficiency with Git and collaborative development workflows (code reviews, pull requests, CI)Strong analytical skills and attention to detail; ability to distinguish signal from noiseExcellent communication skills and ability to clearly explain complex technical findingsCuriosity and commitment to continuous improvement; belief that systems, signals, and processes can always be improvedNice to HaveKnowledge of Linux-based systems, including operating, debugging, and optimizing services on production serversExperience building, deploying, and operating containerized workloads using Docker or similar technologiesExperience working with cloud platforms such as Google Cloud Platform, Amazon Web Services, or Microsoft AzureExperience building or operating real-time or high-volume data pipelines, or working with large-scale internet measurement datasetsExperience with workflow orchestration and scheduling systems such as AirflowFamiliarity with ad tech, fraud detection, abuse prevention, or cybersecurity use casesExperience publishing or presenting technical work, including blog posts, academic papers, whitepapers, or conference talksWhat We OfferBuild at a bootstrapped, independent company with no board or outside investors — we optimize for long-term product quality, not short-term growth targetsReal ownership and autonomy: you’ll shape systems, signals, and direction, not just implement tickets100% remote, globally distributed teamFlexible working hours designed for deep focus and a sustainable paceCompetitive salary, adjusted for experience and local marketFlexible vacation policy built on trust and personal responsibilitySolve hard, real-world problems at internet scale, using data most companies never seeAt least one annual company-wide gathering to reconnect and reset in person.
location, VPN and residential proxy detection, mobile carrier data, and over 20 other context tags.
Our API handles over 120 billion requests per month, and we also license our data for use in many products and services you’ve likely interacted with.
We’re a fast-growing, bootstrapped company with a globally distributed team of around 60 people.
Our data powers customers such as Cloudflare, T-Mobile, SpaceX, DemandBase, and Clearbit, among many others.
We also sponsor and contribute to academic conferences like ACM IMC and ACM CoNEXT, staying closely connected with the internet measurement and research community. About the Privacy TeamThe Privacy team delivers accurate, verifiable insight into how IP addresses are used across the internet.
Instead of relying on opaque risk scores or ambiguous classifications, we focus on defensible, fact-based signals derived from observable network behavior.
Our work centers on identifying and analyzing the applications, protocols, and infrastructure used to anonymize or tunnel internet traffic—such as VPNs and residential proxy networks.
We combine large-scale internet measurement with careful analysis to understand how these technologies are deployed and how they behave in real-world environments. Examples of our research and methodology include:VPN
Location Mismatch Report – examining how VPN exit
locations differ from expected geographies, https://ipinfo. io/blog/vpn-
location-mismatch-reportResidential Proxy Coverage – analyzing the growth and characteristics of residential proxy networks, https://ipinfo. io/blog/residential-proxy-coverage-growsWe collaborate with internet infrastructure providers, including major CDNs and network operators / ISPs, as well as customers across sectors such as banking, ad tech, fraud prevention, and security.
Our work helps these partners understand traffic quality, abuse patterns, and privacy tooling with clarity and confidence. Team members are expected not only to build and maintain systems, but also to challenge assumptions, validate signals, and publish findings that can withstand external scrutiny.
How We WorkWe’re an ambitious, fully remote team spread across the globe.
We sync up on a monthly all-hands Zoom call, and most teams meet every 1–2 weeks. Everything else happens asynchronously using Slack, GitHub, Linear, and Notion.
This setup allows you to choose the hours that work best for you, while still collaborating closely with teammates. Autonomy, ownership, and clear communication are essential to how we operate. ResponsibilitiesDesign, build, and operate data collection and analysis pipelines to detect proxy usage, VPN applications, and related spoofing or evasion behaviorWork with large-scale internet measurement data (we collect 75+ TB per week, including BGP, DNS, ping, and traceroute data from 1200+ global vantage points)Research, apply, and implement techniques from cutting-edge internet measurement and network security researchMaintain a high bar for signal quality and defensibility, prioritizing observable network behavior over heuristics or guessworkCommunicate findings clearly by contributing to blog posts, technical documentation, and research publications, both internally and externallySkills and ExperienceRequiredBackground in one or more of:Internet measurements or network telemetryData engineering or experience with batch, streaming, or real-time data pipelinesNetwork engineering or ISP / CDN operations with experience analyzing traffic behaviorSecurity research or applied network securityThreat intelligence or abuse / fraud analysisDeep understanding of networking protocols and networked applications (e. g. VPNs, proxies, tunneling, routing behavior)Proficiency in Bash scripting, Python, Go, or similar languages for building data collection and analysis systemsProficiency in SQL for querying, analyzing, and validating large datasetsProficiency with Git and collaborative development workflows (code reviews, pull requests, CI)Strong analytical skills and attention to detail; ability to distinguish signal from noiseExcellent communication skills and ability to clearly explain complex technical findingsCuriosity and commitment to continuous improvement; belief that systems, signals, and processes can always be improvedNice to HaveKnowledge of Linux-based systems, including operating, debugging, and optimizing services on production serversExperience building, deploying, and operating containerized workloads using Docker or similar technologiesExperience working with cloud platforms such as Google Cloud Platform, Amazon Web Services, or Microsoft AzureExperience building or operating real-time or high-volume data pipelines, or working with large-scale internet measurement datasetsExperience with workflow orchestration and scheduling systems such as AirflowFamiliarity with ad tech, fraud detection, abuse prevention, or cybersecurity use casesExperience publishing or presenting technical work, including blog posts, academic papers, whitepapers, or conference talksWhat We OfferBuild at a bootstrapped, independent company with no board or outside investors — we optimize for long-term product quality, not short-term growth targetsReal ownership and autonomy: you’ll shape systems, signals, and direction, not just implement tickets100% remote, globally distributed teamFlexible working hours designed for deep focus and a sustainable paceCompetitive salary, adjusted for experience and local marketFlexible vacation policy built on trust and personal responsibilitySolve hard, real-world problems at internet scale, using data most companies never seeAt least one annual company-wide gathering to reconnect and reset in person.
Discover the company
Explore other offers from this company or learn more about IPinfo.
The company
I
IPinfo Remote