Chris Riccomini is a General Partner and seasoned software engineer with 14 years building and scaling data infrastructure from Sunnyvale, California. He combines deep distributed-systems and backend expertise — having helped build LinkedIn’s Samza streaming system and driven BigQuery, Kafka, Debezium, and Airflow adoption at WePay — with hands-on open-source contributions to projects like Apache Airflow and Kafka Connect BigQuery. At WePay he led data and ledger teams, architected offline and streaming pipelines, and shipped durable integrations (including Pandas-BigQuery hooks and schema registry improvements). Beyond engineering, he’s an active seed investor and advisor (early backer of Confluent, Prefect and others) and co‑authored The Missing README, sharing practical lessons for software engineers. Notably, his open-source work includes nuanced improvements such as type-safe and cached schema registries and robust topic-to-table resolution logic, reflecting both pragmatic and detail-oriented craftsmanship.
DEPRECATED. PLEASE USE https://github.com/confluentinc/kafka-connect-bigquery. A Kafka Connect BigQuery sink connector
Role in this project:
Backend & DevOps Engineer
Contributions:3 releases, 27 commits, 60 PRs in 3 years 11 months
Contributions summary:Chris primarily contributed to the codebase by refactoring and improving the topic-to-table resolution logic, including the sanitization of table names. They fixed issues related to Docker integration tests and GCS loader, correcting concurrent modifications and improving performance. Additionally, the user addressed schema update failures by logging table ID information and upgraded the project by applying the 3.1.1 upgrade.
Apache Airflow - A platform to programmatically author, schedule, and monitor workflows
Role in this project:
Back-end & Data Engineer
Contributions:2 reviews, 94 PRs, 22 pushes in 5 years 3 months
Contributions summary:Chris primarily contributed to the integration of BigQuery functionality within the Apache Airflow platform, adding the necessary hooks and operators to facilitate interaction with Google Cloud's data warehouse service. Their work included the creation of a BigQuery hook for interacting with the BigQuery API, adding support for various operations such as running queries, exporting and importing data. The user's contributions also involved integrating Pandas with BigQuery and integrating Google Cloud Storage.
monitorpythonschedulerapacheprogrammatically
Find and Hire Top DevelopersWe’ve analyzed the programming source code of over 60 million software developers on GitHub and scored them by 50,000 skills. Sign-up on Prog,AI to search for software developers.