1 line
No EOL
32 KiB
JSON
1 line
No EOL
32 KiB
JSON
{"slug": "igorbarinov--awesome-data-engineering", "title": "Data Engineering", "description": "A curated list of data engineering tools for software developers", "github_url": "https://github.com/igorbarinov/awesome-data-engineering", "stars": "8.3K", "tag": "Big Data", "entry_count": 170, "subcategory_count": 9, "subcategories": [{"name": "General", "parent": "", "entries": [{"name": "Databases", "url": "#databases", "description": ""}, {"name": "Data Comparison", "url": "#data-comparison", "description": ""}, {"name": "Data Ingestion", "url": "#data-ingestion", "description": ""}, {"name": "File System", "url": "#file-system", "description": ""}, {"name": "Serialization format", "url": "#serialization-format", "description": ""}, {"name": "Stream Processing", "url": "#stream-processing", "description": ""}, {"name": "Batch Processing", "url": "#batch-processing", "description": ""}, {"name": "Charts and Dashboards", "url": "#charts-and-dashboards", "description": ""}, {"name": "Workflow", "url": "#workflow", "description": ""}, {"name": "Data Lake Management", "url": "#data-lake-management", "description": ""}, {"name": "ELK Elastic Logstash Kibana", "url": "#elk-elastic-logstash-kibana", "description": ""}, {"name": "Docker", "url": "#docker", "description": ""}, {"name": "Datasets", "url": "#datasets", "description": ""}, {"name": "Monitoring", "url": "#monitoring", "description": ""}, {"name": "Profiling", "url": "#profiling", "description": ""}, {"name": "Testing", "url": "#testing", "description": ""}, {"name": "Community", "url": "#community", "description": ""}, {"name": "datacompy", "url": "https://github.com/capitalone/datacompy", "description": "A Python library that facilitates the comparison of two DataFrames in Pandas, Polars, Spark and more. The library goes beyond basic equality checks by providing detailed insights into discrepancies at both row and column levels.", "stars": "632"}, {"name": "dvt", "url": "https://github.com/GoogleCloudPlatform/professional-services-data-validator", "description": "Data Validation Tool compares data from source and target tables to ensure that they match. It provides column validation, row validation, schema validation, custom query validation, and ad hoc SQL exploration.", "stars": "493"}, {"name": "koala-diff", "url": "https://github.com/godalida/koala-diff", "description": "A high-performance Python library for comparing large datasets (CSV, Parquet) locally using Rust and Polars. It features zero-copy streaming to prevent OOM errors and generates interactive HTML data quality reports.", "stars": "4"}, {"name": "everyrow", "url": "https://github.com/futuresearch/everyrow-sdk", "description": "AI-powered data operations SDK for Python. Semantic deduplication, fuzzy table merging, and intelligent row ranking using LLM agents.", "stars": "16"}, {"name": "ingestr", "url": "https://github.com/bruin-data/ingestr", "description": "CLI tool to copy data between databases with a single command. Supports 50+ sources including PostgreSQL, MySQL, MongoDB, Salesforce, Shopify to any data warehouse.", "stars": "3.4k"}, {"name": "Kafka", "url": "https://kafka.apache.org/", "description": "Publish-subscribe messaging rethought as a distributed commit log."}, {"name": "AWS Kinesis", "url": "https://aws.amazon.com/kinesis/", "description": "A fully managed, cloud-based service for real-time data processing over large, distributed data streams."}, {"name": "RabbitMQ", "url": "https://www.rabbitmq.com/", "description": "Robust messaging for applications."}, {"name": "dlt", "url": "https://www.dlthub.com", "description": "A fast\\&simple pipeline building library for Python data devs, runs in notebooks, cloud functions, airflow, etc."}, {"name": "FluentD", "url": "https://www.fluentd.org", "description": "An open source data collector for unified logging layer."}, {"name": "Embulk", "url": "https://www.embulk.org", "description": "An open source bulk data loader that helps data transfer between various databases, storages, file formats, and cloud services."}, {"name": "Apache Sqoop", "url": "https://sqoop.apache.org", "description": "A tool designed for efficiently transferring bulk data between Apache Hadoop and structured datastores such as relational databases."}, {"name": "Heka", "url": "https://github.com/mozilla-services/heka", "description": "Data Acquisition and Processing Made Easy. Deprecated.", "stars": "3.5k"}, {"name": "Gobblin", "url": "https://github.com/apache/incubator-gobblin", "description": "Universal data ingestion framework for Hadoop from LinkedIn.", "stars": "2.3k"}, {"name": "Nakadi", "url": "https://nakadi.io", "description": "An open source event messaging platform that provides a REST API on top of Kafka-like queues."}, {"name": "Pravega", "url": "https://www.pravega.io", "description": "Provides a new storage abstraction - a stream - for continuous and unbounded data."}, {"name": "Apache Pulsar", "url": "https://pulsar.apache.org/", "description": "An open-source distributed pub-sub messaging system."}, {"name": "AWS Data Wrangler", "url": "https://github.com/awslabs/aws-data-wrangler", "description": "Utility belt to handle data on AWS.", "stars": "4.1k"}, {"name": "Airbyte", "url": "https://airbyte.io/", "description": "Open-source data integration for modern data teams."}, {"name": "Artie", "url": "https://www.artie.com/", "description": "Real-time data ingestion tool leveraging change data capture."}, {"name": "Sling", "url": "https://slingdata.io/", "description": "CLI data integration tool specialized in moving data between databases, as well as storage systems."}, {"name": "Meltano", "url": "https://meltano.com/", "description": "CLI & code-first ELT."}, {"name": "Google Sheets ETL", "url": "https://github.com/fulldecent/google-sheets-etl", "description": "Live import all your Google Sheets to your data warehouse.", "stars": "22"}, {"name": "CsvPath Framework", "url": "https://www.csvpath.org/", "description": "A delimited data preboarding framework that fills the gap between MFT and the data lake."}, {"name": "Estuary Flow", "url": "https://estuary.dev", "description": "No/low-code data pipeline platform that handles both batch and real-time data ingestion."}, {"name": "db2lake", "url": "https://github.com/bahador-r/db2lake", "description": "Lightweight Node.js ETL framework for databases \u2192 data lakes/warehouses.", "stars": "2"}, {"name": "Kreuzberg", "url": "https://github.com/kreuzberg-dev/kreuzberg", "description": "Polyglot document intelligence library with a Rust core and bindings for Python, TypeScript, Go, and more. Extracts text, tables, and metadata from 62+ document formats for data pipeline ingestion.", "stars": "5.8k"}, {"name": "HDFS", "url": "https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-hdfs/HdfsDesign.html", "description": "A distributed file system designed to run on commodity hardware."}, {"name": "AWS S3", "url": "https://aws.amazon.com/s3/", "description": "Object storage built to retrieve any amount of data from anywhere."}, {"name": "Alluxio", "url": "https://www.alluxio.org/", "description": "A memory-centric distributed storage system enabling reliable data sharing at memory-speed across cluster frameworks, such as Spark and MapReduce."}, {"name": "CEPH", "url": "https://ceph.com/", "description": "A unified, distributed storage system designed for excellent performance, reliability, and scalability."}, {"name": "JuiceFS", "url": "https://github.com/juicedata/juicefs", "description": "A high-performance Cloud-Native file system driven by object storage for large-scale data storage.", "stars": "13k"}, {"name": "OrangeFS", "url": "https://www.orangefs.org/", "description": "Orange File System is a branch of the Parallel Virtual File System."}, {"name": "SnackFS", "url": "https://github.com/tuplejump/snackfs-release", "description": "A bite-sized, lightweight HDFS compatible file system built over Cassandra.", "stars": "14"}, {"name": "GlusterFS", "url": "https://www.gluster.org/", "description": "Gluster Filesystem."}, {"name": "XtreemFS", "url": "https://www.xtreemfs.org/", "description": "Fault-tolerant distributed file system for all storage needs."}, {"name": "SeaweedFS", "url": "https://github.com/chrislusf/seaweedfs", "description": "Seaweed-FS is a simple and highly scalable distributed file system. There are two objectives: to store billions of files! to serve the files fast! Instead of supporting full POSIX file system semantics, Seaweed-FS choose to implement only a key\\~file mapping. Similar to the word \"NoSQL\", you can call it as \"NoFS\".", "stars": "16"}, {"name": "S3QL", "url": "https://github.com/s3ql/s3ql/", "description": "A file system that stores all its data online using storage services like Google Storage, Amazon S3, or OpenStack.", "stars": "1.2k"}, {"name": "LizardFS", "url": "https://lizardfs.com/", "description": "Software Defined Storage is a distributed, parallel, scalable, fault-tolerant, Geo-Redundant and highly available file system."}, {"name": "Apache Avro", "url": "https://avro.apache.org", "description": "Apache Avro\u2122 is a data serialization system."}, {"name": "Apache Parquet", "url": "https://parquet.apache.org", "description": "A columnar storage format available to any project in the Hadoop ecosystem, regardless of the choice of data processing framework, data model or programming language."}, {"name": "Apache ORC", "url": "https://orc.apache.org/", "description": "The smallest, fastest columnar storage for Hadoop workloads."}, {"name": "Apache Thrift", "url": "https://thrift.apache.org", "description": "The Apache Thrift software framework, for scalable cross-language services development."}, {"name": "ProtoBuf", "url": "https://github.com/protocolbuffers/protobuf", "description": "Protocol Buffers - Google's data interchange format.", "stars": "71k"}, {"name": "SequenceFile", "url": "https://wiki.apache.org/hadoop/SequenceFile", "description": "A flat file consisting of binary key/value pairs. It is extensively used in MapReduce as input/output formats."}, {"name": "Kryo", "url": "https://github.com/EsotericSoftware/kryo", "description": "A fast and efficient object graph serialization framework for Java.", "stars": "6.5k"}, {"name": "Apache Beam", "url": "https://beam.apache.org/", "description": "A unified programming model that implements both batch and streaming data processing jobs that run on many execution engines."}, {"name": "Spark Streaming", "url": "https://spark.apache.org/streaming/", "description": "Makes it easy to build scalable fault-tolerant streaming applications."}, {"name": "Apache Flink", "url": "https://flink.apache.org/", "description": "A streaming dataflow engine that provides data distribution, communication, and fault tolerance for distributed computations over data streams."}, {"name": "Apache Storm", "url": "https://storm.apache.org", "description": "A free and open source distributed realtime computation system."}, {"name": "Apache Samza", "url": "https://samza.apache.org", "description": "A distributed stream processing framework."}, {"name": "Apache NiFi", "url": "https://nifi.apache.org/", "description": "An easy to use, powerful, and reliable system to process and distribute data."}, {"name": "Apache Hudi", "url": "https://hudi.apache.org/", "description": "An open source framework for managing storage for real time processing, one of the most interesting feature is the Upsert."}, {"name": "CocoIndex", "url": "https://github.com/cocoindex-io/cocoindex", "description": "An open source ETL framework to build fresh index for AI.", "stars": "6.1k"}, {"name": "VoltDB", "url": "https://voltdb.com/", "description": "An ACID-compliant RDBMS which uses a [shared nothing architecture](https://en.wikipedia.org/wiki/Shared-nothing_architecture)."}, {"name": "PipelineDB", "url": "https://github.com/pipelinedb/pipelinedb", "description": "The Streaming SQL Database.", "stars": "2.7k"}, {"name": "Spring Cloud Dataflow", "url": "https://cloud.spring.io/spring-cloud-dataflow/", "description": "Streaming and tasks execution between Spring Boot apps."}, {"name": "Bonobo", "url": "https://www.bonobo-project.org/", "description": "A data-processing toolkit for python 3.5+."}, {"name": "Robinhood's Faust", "url": "https://github.com/faust-streaming/faust", "description": "Forever scalable event processing & in-memory durable K/V store as a library with asyncio & static typing.", "stars": "1.9k"}, {"name": "HStreamDB", "url": "https://github.com/hstreamdb/hstream", "description": "The streaming database built for IoT data storage and real-time processing.", "stars": "726"}, {"name": "Kuiper", "url": "https://github.com/emqx/kuiper", "description": "An edge lightweight IoT data analytics/streaming software implemented by Golang, and it can be run at all kinds of resource-constrained edge devices.", "stars": "1.7k"}, {"name": "Zilla", "url": "https://github.com/aklivity/zilla", "description": "- An API gateway built for event-driven architectures and streaming that supports standard protocols such as HTTP, SSE, gRPC, MQTT, and the native Kafka protocol.", "stars": "678"}, {"name": "SwimOS", "url": "https://github.com/swimos/swim-rust", "description": "A framework for building real-time streaming data processing applications that supports a wide range of ingestion sources."}, {"name": "Pathway", "url": "https://github.com/pathwaycom/pathway", "description": "Performant open-source Python ETL framework with Rust runtime, supporting 300+ data sources.", "stars": "59k"}, {"name": "Hadoop MapReduce", "url": "https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html", "description": "A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) - in-parallel on large clusters (thousands of nodes) - of commodity hardware in a reliable, fault-tolerant manner."}, {"name": "Spark", "url": "https://spark.apache.org/", "description": "A multi-language engine for executing data engineering, data science, and machine learning on single-node machines or clusters."}, {"name": "AWS EMR", "url": "https://aws.amazon.com/emr/", "description": "A web service that makes it easy to quickly and cost-effectively process vast amounts of data."}, {"name": "Data Mechanics", "url": "https://www.datamechanics.co", "description": "A cloud-based platform deployed on Kubernetes making Apache Spark more developer-friendly and cost-effective."}, {"name": "Tez", "url": "https://tez.apache.org/", "description": "An application framework which allows for a complex directed-acyclic-graph of tasks for processing data."}, {"name": "Bistro", "url": "https://github.com/asavinov/bistro", "description": "A light-weight engine for general-purpose data processing including both batch and stream analytics. It is based on a novel unique data model, which represents data via *functions* and processes data via *columns operations* as opposed to having only set operations in conventional approaches like MapReduce or SQL.", "stars": "8"}, {"name": "Substation", "url": "https://github.com/brexhq/substation", "description": "A cloud native data pipeline and transformation toolkit written in Go.", "stars": "389"}, {"name": "Highcharts", "url": "https://www.highcharts.com/", "description": "A charting library written in pure JavaScript, offering an easy way of adding interactive charts to your web site or web application."}, {"name": "ZingChart", "url": "https://www.zingchart.com/", "description": "Fast JavaScript charts for any data set."}, {"name": "C3.js", "url": "https://c3js.org", "description": "D3-based reusable chart library."}, {"name": "D3.js", "url": "https://d3js.org/", "description": "A JavaScript library for manipulating documents based on data."}, {"name": "SmoothieCharts", "url": "https://smoothiecharts.org", "description": "A JavaScript Charting Library for Streaming Data."}, {"name": "PyXley", "url": "https://github.com/stitchfix/pyxley", "description": "Python helpers for building dashboards using Flask and React.", "stars": "2.3k"}, {"name": "Plotly", "url": "https://github.com/plotly/dash", "description": "Flask, JS, and CSS boilerplate for interactive, web-based visualization apps in Python.", "stars": "25k"}, {"name": "Apache Superset", "url": "https://github.com/apache/incubator-superset", "description": "A modern, enterprise-ready business intelligence web application.", "stars": "71k"}, {"name": "Redash", "url": "https://redash.io/", "description": "Make Your Company Data Driven. Connect to any data source, easily visualize and share your data."}, {"name": "Metabase", "url": "https://github.com/metabase/metabase", "description": "The easy, open source way for everyone in your company to ask questions and learn from data.", "stars": "46k"}, {"name": "PyQtGraph", "url": "https://www.pyqtgraph.org/", "description": "A pure-python graphics and GUI library built on PyQt4 / PySide and numpy. It is intended for use in mathematics / scientific / engineering applications."}, {"name": "Seaborn", "url": "https://seaborn.pydata.org", "description": "A Python visualization library based on matplotlib. It provides a high-level interface for drawing attractive statistical graphics."}, {"name": "QueryGPT", "url": "https://github.com/MKY508/QueryGPT", "description": "Natural language database query interface with automatic chart generation, supporting Chinese and English queries.", "stars": "28"}, {"name": "Bruin", "url": "https://github.com/bruin-data/bruin", "description": "End-to-end data pipeline tool that combines ingestion, transformation (SQL + Python), and data quality in a single CLI. Connects to BigQuery, Snowflake, PostgreSQL, Redshift, and more. Includes VS Code extension with live previews.", "stars": "1.4k"}, {"name": "Luigi", "url": "https://github.com/spotify/luigi", "description": "A Python module that helps you build complex pipelines of batch jobs.", "stars": "19k"}, {"name": "CronQ", "url": "https://github.com/seatgeek/cronq", "description": "An application cron-like system. [Used](https://chairnerd.seatgeek.com/building-out-the-seatgeek-data-pipeline/) w/Luigi. Deprecated."}, {"name": "Cascading", "url": "https://www.cascading.org/", "description": "Java based application development platform."}, {"name": "Airflow", "url": "https://github.com/apache/airflow", "description": "A system to programmatically author, schedule, and monitor data pipelines.", "stars": "44k"}, {"name": "Azkaban", "url": "https://azkaban.github.io/", "description": "A batch workflow job scheduler created at LinkedIn to run Hadoop jobs. Azkaban resolves the ordering through job dependencies and provides an easy-to-use web user interface to maintain and track your workflows."}, {"name": "Oozie", "url": "https://oozie.apache.org/", "description": "A workflow scheduler system to manage Apache Hadoop jobs."}, {"name": "Pinball", "url": "https://github.com/pinterest/pinball", "description": "DAG based workflow manager. Job flows are defined programmatically in Python. Support output passing between jobs.", "stars": "1k"}, {"name": "Dagster", "url": "https://github.com/dagster-io/dagster", "description": "An open-source Python library for building data applications.", "stars": "15k"}, {"name": "Hamilton", "url": "https://github.com/dagworks-inc/hamilton", "description": "A lightweight library to define data transformations as a directed-acyclic graph (DAG). If you like dbt for SQL transforms, you will like Hamilton for Python processing.", "stars": "2.4k"}, {"name": "Kedro", "url": "https://kedro.readthedocs.io/en/latest/", "description": "A framework that makes it easy to build robust and scalable data pipelines by providing uniform project templates, data abstraction, configuration and pipeline assembly."}, {"name": "Dataform", "url": "https://dataform.co/", "description": "An open-source framework and web based IDE to manage datasets and their dependencies. SQLX extends your existing SQL warehouse dialect to add features that support dependency management, testing, documentation and more."}, {"name": "Census", "url": "https://getcensus.com/", "description": "A reverse-ETL tool that let you sync data from your cloud data warehouse to SaaS applications like Salesforce, Marketo, HubSpot, Zendesk, etc. No engineering favors required\u2014just SQL."}, {"name": "dbt", "url": "https://getdbt.com/", "description": "A command line tool that enables data analysts and engineers to transform data in their warehouses more effectively."}, {"name": "Kestra", "url": "https://github.com/kestra-io/kestra", "description": "Scalable, event-driven, language-agnostic orchestration and scheduling platform to manage millions of workflows declaratively in code.", "stars": "26k"}, {"name": "RudderStack", "url": "https://github.com/rudderlabs/rudder-server", "description": "A warehouse-first Customer Data Platform that enables you to collect data from every application, website and SaaS platform, and then activate it in your warehouse and business tools.", "stars": "4.4k"}, {"name": "PACE", "url": "https://github.com/getstrm/pace", "description": "An open source framework that allows you to enforce agreements on how data should be accessed, used, and transformed, regardless of the data platform (Snowflake, BigQuery, DataBricks, etc.)", "stars": "38"}, {"name": "Prefect", "url": "https://prefect.io/", "description": "An orchestration and observability platform. With it, developers can rapidly build and scale resilient code, and triage disruptions effortlessly."}, {"name": "Multiwoven", "url": "https://github.com/Multiwoven/multiwoven", "description": "The open-source reverse ETL, data activation platform for modern data teams.", "stars": "1.6k"}, {"name": "SuprSend", "url": "https://www.suprsend.com/products/workflows", "description": "Create automated workflows and logic using API's for your notification service. Add templates, batching, preferences, inapp inbox with workflows to trigger notifications directly from your data warehouse."}, {"name": "Mage", "url": "https://www.mage.ai", "description": "Open-source data pipeline tool for transforming and integrating data."}, {"name": "SQLMesh", "url": "https://sqlmesh.readthedocs.io", "description": "An open-source data transformation framework for managing, testing, and deploying SQL and Python-based data pipelines with version control, environment isolation, and automatic dependency resolution."}, {"name": "lakeFS", "url": "https://github.com/treeverse/lakeFS", "description": "An open source platform that delivers resilience and manageability to object-storage based data lakes.", "stars": "5.2k"}, {"name": "Project Nessie", "url": "https://github.com/projectnessie/nessie", "description": "A Transactional Catalog for Data Lakes with Git-like semantics. Works with Apache Iceberg tables.", "stars": "1.4k"}, {"name": "Ilum", "url": "https://ilum.cloud/", "description": "A modular Data Lakehouse platform that simplifies the management and monitoring of Apache Spark clusters across Kubernetes and Hadoop environments."}, {"name": "Gravitino", "url": "https://github.com/apache/gravitino", "description": "An open-source, unified metadata management for data lakes, data warehouses, and external catalogs.", "stars": "2.9k"}, {"name": "FlightPath Data", "url": "https://www.flightpathdata.com", "description": "FlightPath is a gateway to a data lake's bronze layer, protecting it from invalid external data file feeds as a trusted publisher."}, {"name": "docker-logstash", "url": "https://github.com/pblittle/docker-logstash", "description": "A highly configurable Logstash (1.4.4) - Docker image running Elasticsearch (1.7.0) - and Kibana (3.1.2).", "stars": "237"}, {"name": "elasticsearch-jdbc", "url": "https://github.com/jprante/elasticsearch-jdbc", "description": "JDBC importer for Elasticsearch.", "stars": "2.8k"}, {"name": "ZomboDB", "url": "https://github.com/zombodb/zombodb", "description": "PostgreSQL Extension that allows creating an index backed by Elasticsearch.", "stars": "4.7k"}, {"name": "Gockerize", "url": "https://github.com/redbooth/gockerize", "description": "Package golang service into minimal Docker containers.", "stars": "667"}, {"name": "Flocker", "url": "https://github.com/ClusterHQ/flocker", "description": "Easily manage Docker containers & their data.", "stars": "3.4k"}, {"name": "Rancher", "url": "https://rancher.com/rancher-os/", "description": "RancherOS is a 20mb Linux distro that runs the entire OS as Docker containers."}, {"name": "Kontena", "url": "https://www.kontena.io/", "description": "Application Containers for Masses."}, {"name": "Weave", "url": "https://github.com/weaveworks/weave", "description": "Weaving Docker containers into applications.", "stars": "6.6k"}, {"name": "Zodiac", "url": "https://github.com/CenturyLinkLabs/zodiac", "description": "A lightweight tool for easy deployment and rollback of dockerized applications.", "stars": "200"}, {"name": "cAdvisor", "url": "https://github.com/google/cadvisor", "description": "Analyzes resource usage and performance characteristics of running containers.", "stars": "19k"}, {"name": "Micro S3 persistence", "url": "https://github.com/figadore/micro-s3-persistence", "description": "Docker microservice for saving/restoring volume data to S3.", "stars": "14"}, {"name": "Rocker-compose", "url": "https://github.com/grammarly/rocker-compose", "description": "Docker composition tool with idempotency features for deploying apps composed of multiple containers. Deprecated.", "stars": "408"}, {"name": "Nomad", "url": "https://github.com/hashicorp/nomad", "description": "A cluster manager, designed for both long-lived services and short-lived batch processing workloads.", "stars": "16k"}, {"name": "ImageLayers", "url": "https://imagelayers.io/", "description": "Visualize Docker images and the layers that compose them."}]}, {"name": "Realtime", "parent": "Datasets", "entries": [{"name": "Twitter Realtime", "url": "https://developer.twitter.com/en/docs/tweets/filter-realtime/overview", "description": "The Streaming APIs give developers low latency access to Twitter's global stream of Tweet data."}, {"name": "Eventsim", "url": "https://github.com/Interana/eventsim", "description": "Event data simulator. Generates a stream of pseudo-random events from a set of users, designed to simulate web traffic.", "stars": "535"}, {"name": "Reddit", "url": "https://www.reddit.com/r/datasets/comments/3mk1vg/realtime_data_is_available_including_comments/", "description": "Real-time data is available including comments, submissions and links posted to reddit."}]}, {"name": "Data Dumps", "parent": "Datasets", "entries": [{"name": "GitHub Archive", "url": "https://www.gharchive.org/", "description": "GitHub's public timeline since 2011, updated every hour."}, {"name": "Common Crawl", "url": "https://commoncrawl.org/", "description": "Open source repository of web crawl data."}, {"name": "Wikipedia", "url": "https://dumps.wikimedia.org/enwiki/latest/", "description": "Wikipedia's complete copy of all wikis, in the form of Wikitext source and metadata embedded in XML. A number of raw database tables in SQL form are also available."}]}, {"name": "Prometheus", "parent": "Monitoring", "entries": [{"name": "Prometheus.io", "url": "https://github.com/prometheus/prometheus", "description": "An open-source service monitoring system and time series database.", "stars": "63k"}, {"name": "HAProxy Exporter", "url": "https://github.com/prometheus/haproxy_exporter", "description": "Simple server that scrapes HAProxy stats and exports them via HTTP for Prometheus consumption.", "stars": "627"}]}, {"name": "Data Profiler", "parent": "Profiling", "entries": [{"name": "Data Profiler", "url": "https://github.com/capitalone/dataprofiler", "description": "The DataProfiler is a Python library designed to make data analysis, monitoring, and sensitive data detection easy.", "stars": "1.5k"}, {"name": "YData Profiling", "url": "https://docs.profiling.ydata.ai/latest/", "description": "A general-purpose open-source data profiler for high-level analysis of a dataset."}, {"name": "Desbordante", "url": "https://github.com/desbordante/desbordante-core", "description": "An open-source data profiler specifically focused on discovery and validation of complex patterns in data.", "stars": "465"}, {"name": "Grai", "url": "https://github.com/grai-io/grai-core/", "description": "A data catalog tool that integrates into your CI system exposing downstream impact testing of data changes. These tests prevent data changes which might break data pipelines or BI dashboards from making it to production.", "stars": "313"}, {"name": "DQOps", "url": "https://github.com/dqops/dqo", "description": "An open-source data quality platform for the whole data platform lifecycle from profiling new data sources to applying full automation of data quality monitoring.", "stars": "187"}, {"name": "DataKitchen", "url": "https://datakitchen.io/", "description": "Open Source Data Observability for end-to-end Data Journey Observability, data profiling, anomaly detection, and auto-created data quality validation tests."}, {"name": "GreatExpectation", "url": "https://greatexpectations.io/", "description": "Open Source data validation framework to manage data quality. Users can define and document \u201cexpectations\u201d rules about how data should look and behave."}, {"name": "RunSQL", "url": "https://runsql.com/", "description": "Free online SQL playground for MySQL, PostgreSQL, and SQL Server. Create database structures, run queries, and share results instantly."}, {"name": "Spark Playground", "url": "https://www.sparkplayground.com/", "description": "Write, run, and test PySpark code on Spark Playground's online compiler. Access real-world sample datasets & solve interview questions to enhance your PySpark skills for data engineering roles."}, {"name": "daffy", "url": "https://github.com/vertti/daffy/", "description": "Decorator-first DataFrame contracts/validation (columns/dtypes/constraints) at function boundaries. Supports Pandas/Polars/PyArrow/Modin.", "stars": "53"}, {"name": "Snowflake Emulator", "url": "https://github.com/nnnkkk7/snowflake-emulator", "description": "A Snowflake-compatible emulator for local development and testing.", "stars": "24"}]}, {"name": "Forums", "parent": "Community", "entries": [{"name": "/r/dataengineering", "url": "https://www.reddit.com/r/dataengineering/", "description": "News, tips, and background on Data Engineering."}, {"name": "/r/etl", "url": "https://www.reddit.com/r/ETL/", "description": "Subreddit focused on ETL."}]}, {"name": "Conferences", "parent": "Community", "entries": [{"name": "Data Council", "url": "https://www.datacouncil.ai/about", "description": "The first technical conference that bridges the gap between data scientists, data engineers and data analysts."}]}, {"name": "Podcasts", "parent": "Community", "entries": [{"name": "Data Engineering Podcast", "url": "https://www.dataengineeringpodcast.com/", "description": "The show about modern data infrastructure."}, {"name": "The Data Stack Show", "url": "https://datastackshow.com/", "description": "A show where they talk to data engineers, analysts, and data scientists about their experience around building and maintaining data infrastructure, delivering data and data products, and driving better outcomes across their businesses with data."}]}, {"name": "Books", "parent": "Community", "entries": [{"name": "Snowflake Data Engineering", "url": "https://www.manning.com/books/snowflake-data-engineering", "description": "A practical introduction to data engineering on the Snowflake cloud data platform."}, {"name": "Best Data Science Books", "url": "https://www.appliedaicourse.com/blog/data-science-books/", "description": "This blog offers a curated list of top data science books, categorized by topics and learning stages, to aid readers in building foundational knowledge and staying updated with industry trends."}, {"name": "Architecting an Apache Iceberg Lakehouse", "url": "https://www.manning.com/books/architecting-an-apache-iceberg-lakehouse", "description": "A guide to designing an Apache Iceberg lakehouse from scratch."}, {"name": "Learn AI Data Engineering in a Month of Lunches", "url": "https://www.manning.com/books/learn-ai-data-engineering-in-a-month-of-lunches", "description": "A fast, friendly guide to integrating large language models into your data workflows."}]}], "name": ""} |