Unlocking Competitive Edge with Big Data, Data Sources & Databases

About Course
About This Course
Learners will be able to:
• Develop efficient processes to standardise and maintain data definitions, sources and quality
• Develop data warehouse process models, including sourcing, loading, transformation and extraction
• Design data validation methodology to verify reliability and validity of data
• Design staging databases to store the data temporarily before moving them into the target system
• Design extraction process for consolidating data from multiple data source systems
• Verify extracted data with business rules specified in target system
• Design the process to transform extracted data into structures that align to the business rules incorporated in the target system
• Develop load process to upload transformed and integrated data to live target system
• Translate complex functional and technical business requirements into detailed data structures and designs
• Develop data integration procedures, managing the alignment of data availability and integration processes
What You’ll Learn
This unit will equip learners with the necessary skills to work with big data, different data sources and databases, storage and processing technologies like Hadoop and Spark, distributed computing with Hadoop and Spark, data warehousing and ETL, NoSQL databases like MongoDB and Cassandra, and stream processing with Kafka and Storm. It also covers an overview and types of data sources and databases, data modelling and normalization, structured data sources and databases, relational databases, SQL queries, MySQL and PostgreSQL installation, configuration and usage, and data warehousing and OLAP.