CodeOp’s premier 6-week Big Data Engineering bootcamp is designed to advance the careers of women who are either working or studying in a technical field. In this course, students learn about the technologies behind the data management and data quality needed to support an organization’s data science workflow.
Beginning first with a clear understanding of the use and limitations of traditional data warehousing tools, students learn when and how to apply the latest open-source technologies used within various Cloud frameworks. Students build data pipelines, stitch data together using scripting languages, work with DBA to construct data stores as well as learn how to scale machine learning projects. The technologies used in this course include the Hadoop Distributed File System, Docker, MongoDB, Google Cloud’s Big Query, AWS Redshift, Spark, Kafka, Beam and Airflow.
Students should expect to have gained knowledge of the major frameworks and technologies used to support a company’s data science workflow, including the Hadoop Distributed File System, Docker, MongoDB, Google Cloud’s Big Query, AWS Redshift, Spark, Kafka, Beam and Airflow. They will have a portfolio of projects which will include building ETLs from scratch and live streaming data using Kafka.