![tuplespectra](/img/default-banner.jpg)
- 77
- 226 511
tuplespectra
เข้าร่วมเมื่อ 31 ธ.ค. 2022
Process Engineer to Machine Learning Engineer | Interview with Raphael | Exponential Salary Increase
#datascientist #motivationalstory #inspiry story
Listen to Raphael sharing his journey on how he made the career switch from Process Engineering into Machine Learning Engineering.
*Books I recommend*
1. Fundamentals of Data Engineering: Plan and Build Robust Data Systems: amzn.to/4bfzmvX
2. Grit: The Power of Passion and Perseverance amzn.to/3EZKSgb
3. Think and Grow Rich!: The Original Version, Restored and Revised: amzn.to/3Q2K68s
4. The Book on Rental Property Investing: How to Create Wealth With Intelligent Buy and Hold Real Estate Investing: amzn.to/3LLpXRy
5. How to Invest in Real Estate: The Ultimate Beginner's Guide to Getting Started: amzn.to/48RbuOb
6. Introducing Python: Modern Computing in Simple Packages amzn.to/3Q4driR
7. Python for Data Analysis: Data Wrangling with pandas, NumPy, and Jupyter 3rd Edition: amzn.to/3rGF73G
Please don’t forget to like, share, comment and subscribe to our channel for more videos.
*My gadgets*
1. Sony ZV-1 II Vlog Camera amzn.to/48Uwoey
2. Carrying & Protective Case for Digital Camera amzn.to/3vWn4IJ
3. WISAMIC Ring Light Kit: 18" 55W LED Ring Light amzn.to/3ue6iUV
4. KHOMO GEAR Extra Large 77 x 62 inch Wide Collapsible amzn.to/3SD6pCJ
DISCLAIMER: This video and description has affiliate links. This means when you buy through one of these links, we will receive a small commission and this is at no cost to you. This will help support us to continue making awesome and valuable contents for you.
Listen to Raphael sharing his journey on how he made the career switch from Process Engineering into Machine Learning Engineering.
*Books I recommend*
1. Fundamentals of Data Engineering: Plan and Build Robust Data Systems: amzn.to/4bfzmvX
2. Grit: The Power of Passion and Perseverance amzn.to/3EZKSgb
3. Think and Grow Rich!: The Original Version, Restored and Revised: amzn.to/3Q2K68s
4. The Book on Rental Property Investing: How to Create Wealth With Intelligent Buy and Hold Real Estate Investing: amzn.to/3LLpXRy
5. How to Invest in Real Estate: The Ultimate Beginner's Guide to Getting Started: amzn.to/48RbuOb
6. Introducing Python: Modern Computing in Simple Packages amzn.to/3Q4driR
7. Python for Data Analysis: Data Wrangling with pandas, NumPy, and Jupyter 3rd Edition: amzn.to/3rGF73G
Please don’t forget to like, share, comment and subscribe to our channel for more videos.
*My gadgets*
1. Sony ZV-1 II Vlog Camera amzn.to/48Uwoey
2. Carrying & Protective Case for Digital Camera amzn.to/3vWn4IJ
3. WISAMIC Ring Light Kit: 18" 55W LED Ring Light amzn.to/3ue6iUV
4. KHOMO GEAR Extra Large 77 x 62 inch Wide Collapsible amzn.to/3SD6pCJ
DISCLAIMER: This video and description has affiliate links. This means when you buy through one of these links, we will receive a small commission and this is at no cost to you. This will help support us to continue making awesome and valuable contents for you.
มุมมอง: 402
วีดีโอ
Petroleum Engineer to Data scientist | Join Africans in Data Community to expand your network
มุมมอง 3613 หลายเดือนก่อน
#datascientist #motivationalstory #communityservice #buildyournetwork In this video, Toyosi Bamidele discussed how she transitioned from a petroleum engineer to a data scientist and the motivation behind founding “Africans in Data Community”. She further explained how you can become a member of this community and the benefits you stand to gain when you become a member. In addition, she gave som...
Complete Roadmap/Guide on How to Become a Data Engineer in 2024
มุมมอง 8654 หลายเดือนก่อน
#dataengineering #dataengineeringroadmap In this video, I share with you a complete guide/roadmap on how to become a data engineer in 2024. You do not have to know everything highlighted in this guide before you start applying for DE roles. Please click on the google drive link below to request a view access to the google doc where I listed the roadmap. docs.google.com/document/d/1hWkKlPvZOBigm...
Tips to become a data engineer in 2024 - What no one ever told you | Keep Learning | Be Consistent
มุมมอง 5695 หลายเดือนก่อน
In this video, I will like to share with you what I have learnt while learning programming and tips to become a data engineer in 2024. *Books I recommend* 1. Grit: The Power of Passion and Perseverance amzn.to/3EZKSgb 2. Think and Grow Rich!: The Original Version, Restored and Revised: amzn.to/3Q2K68s 3. The Book on Rental Property Investing: How to Create Wealth With Intelligent Buy and Hold R...
Tutorial 3.1-How to load/restore DVD rental database to PostgreSQL server16 using pgAdmin| Windows11
มุมมอง 3.2K6 หลายเดือนก่อน
In this video I showed how to load or restore the DVD rental database to our PostgreSQL sever 16 on Windows 11 computer. We will use this DVD rental database for the next series of videos to understand the various SQL concepts from the beginner level to intermediate/advanced level. Please go to this link to download the dvdrental.tar sample database file www.postgresqltutorial.com/postgresql-ge...
Tutorial 2.1 - How to install PostgreSQL 16 and pgAdmin on windows 11 computer | Step-by-Step Guide
มุมมอง 1.1K6 หลายเดือนก่อน
I will be showing you a step by step guide on how to download and install PostgreSQL server 16 and PgAdmin on your windows 11 computer. Follow this link to download postgreSQL www.postgresql.org/download/ *Books I recommend* 1. Grit: The Power of Passion and Perseverance amzn.to/3EZKSgb 2. Think and Grow Rich!: The Original Version, Restored and Revised: amzn.to/3Q2K68s 3. The Book on Rental Pr...
Master SQL | Go from Zero to Hero | SQL Live Class BootCamp
มุมมอง 9906 หลายเดือนก่อน
I am excited to launch this 8-weekends intensive SQL bootcamp which will cover Basic, Intermediate and Advance SQL concepts. 👉 Training start date: Jan 27th 2024. 👉 Days: Saturday and Sunday 👉 Class Time: 5am PST | 6:30pm IST | 2pm WAT 👉 Fee: 40USD 👉 Certificate of Completion will be provided 👉 Register for BootCamp here: awesumedge.com/register-jan2024postgresql/ PostgreSQL Playlist: th-cam.co...
Airflow Tutorial | Automate EMR ETL Jobs with Airflow | Airflow Project | Data Engineering Project
มุมมอง 1.9K6 หลายเดือนก่อน
#dataengineering #emr #airflow #spark #pyspark #aws #etlpipeline #redfin In this video, I explained how to use airflow to automate EMR jobs. I showed you how to create EMR cluster, poll the states of the EMR, add EMR steps and terminate the EMR cluster. Airflow was used to orchestrate the entire ETL pipeline. The EMR steps involved extracting redfin data from the redfin datacenter web address a...
Part 7 - Terminate EMR cluster task | Airflow Tutorial | Automate EMR Jobs with Airflow
มุมมอง 3836 หลายเดือนก่อน
#dataengineering #emr #airflow #spark #pyspark #aws #etlpipeline #redfin In this video, I explained how to use airflow to automate EMR jobs. I showed you how to create EMR cluster, poll the states of the EMR, add EMR steps and terminate the EMR cluster. Airflow was used to orchestrate the entire ETL pipeline. The EMR steps involved extracting redfin data from the redfin datacenter web address a...
Part 6 - Create EMR cluster and Add steps tasks | Airflow Tutorial | Automate EMR Jobs with Airflow
มุมมอง 1.2K6 หลายเดือนก่อน
#dataengineering #emr #airflow #spark #pyspark #aws #etlpipeline #redfin In this video, I explained how to use airflow to automate EMR jobs. I showed you how to create EMR cluster, poll the states of the EMR, add EMR steps and terminate the EMR cluster. Airflow was used to orchestrate the entire ETL pipeline. The EMR steps involved extracting redfin data from the redfin datacenter web address a...
Part 1 - Project Overview | Airflow Tutorial | Automate EMR Jobs with Airflow
มุมมอง 1.4K7 หลายเดือนก่อน
#dataengineering #emr #airflow #spark #pyspark #aws #etlpipeline #redfin In this video, I explained how to use airflow to automate EMR jobs. I showed you how to create EMR cluster, poll the states of the EMR, add EMR steps and terminate the EMR cluster. Airflow was used to orchestrate the entire ETL pipeline. The EMR steps involved extracting redfin data from the redfin datacenter web address a...
Part 3 - Initialize airflow and explore UI | Airflow Tutorial | Automate EMR Jobs with Airflow
มุมมอง 3637 หลายเดือนก่อน
#dataengineering #emr #airflow #spark #pyspark #aws #etlpipeline #redfin In this video, I explained how to use airflow to automate EMR jobs. I showed you how to create EMR cluster, poll the states of the EMR, add EMR steps and terminate the EMR cluster. Airflow was used to orchestrate the entire ETL pipeline. The EMR steps involved extracting redfin data from the redfin datacenter web address a...
Part 5 - Connect VS code and create task 1 | Airflow Tutorial | Automate EMR Jobs with Airflow
มุมมอง 4817 หลายเดือนก่อน
#dataengineering #emr #airflow #spark #pyspark #aws #etlpipeline #redfin In this video, I explained how to use airflow to automate EMR jobs. I showed you how to create EMR cluster, poll the states of the EMR, add EMR steps and terminate the EMR cluster. Airflow was used to orchestrate the entire ETL pipeline. The EMR steps involved extracting redfin data from the redfin datacenter web address a...
Part 4 - Create S3 buckets and required folders | Airflow Tutorial | Automate EMR Jobs with Airflow
มุมมอง 2857 หลายเดือนก่อน
#dataengineering #emr #airflow #spark #pyspark #aws #etlpipeline #redfin In this video, I explained how to use airflow to automate EMR jobs. I showed you how to create EMR cluster, poll the states of the EMR, add EMR steps and terminate the EMR cluster. Airflow was used to orchestrate the entire ETL pipeline. The EMR steps involved extracting redfin data from the redfin datacenter web address a...
Part 2 - Provision EC2 and Install dependencies | Airflow Tutorial | Automate EMR Jobs with Airflow
มุมมอง 6677 หลายเดือนก่อน
#dataengineering #emr #airflow #spark #pyspark #aws #etlpipeline #redfin In this video, I explained how to use airflow to automate EMR jobs. I showed you how to create EMR cluster, poll the states of the EMR, add EMR steps and terminate the EMR cluster. Airflow was used to orchestrate the entire ETL pipeline. The EMR steps involved extracting redfin data from the redfin datacenter web address a...
Part 6 - Data Transformation|pyspark|What is AWS EMR| Extract and Transform Redfin data with AWS EMR
มุมมอง 5657 หลายเดือนก่อน
Part 6 - Data Transformation|pyspark|What is AWS EMR| Extract and Transform Redfin data with AWS EMR
Part 5 - EMR Studio|Jupyter Notebook|What is AWS EMR| Extract and Transform Redfin data with AWS EMR
มุมมอง 3857 หลายเดือนก่อน
Part 5 - EMR Studio|Jupyter Notebook|What is AWS EMR| Extract and Transform Redfin data with AWS EMR
Part 4 - How to Create EMR Cluster | What is AWS EMR| Extract and Transform Redfin data with AWS EMR
มุมมอง 4147 หลายเดือนก่อน
Part 4 - How to Create EMR Cluster | What is AWS EMR| Extract and Transform Redfin data with AWS EMR
Part 3 - How to create AWS VPC | What is AWS EMR| Extract and Transform Redfin data with AWS EMR
มุมมอง 2557 หลายเดือนก่อน
Part 3 - How to create AWS VPC | What is AWS EMR| Extract and Transform Redfin data with AWS EMR
Part 2 - Create User and S3 bucket | What is AWS EMR| Extract and Transform Redfin data with AWS EMR
มุมมอง 3347 หลายเดือนก่อน
Part 2 - Create User and S3 bucket | What is AWS EMR| Extract and Transform Redfin data with AWS EMR
Part 1 - Project Overview | What is AWS EMR | Extract and Transform Redfin data with AWS EMR
มุมมอง 7877 หลายเดือนก่อน
Part 1 - Project Overview | What is AWS EMR | Extract and Transform Redfin data with AWS EMR
What is AWS EMR | Extract and Transform Redfin data with AWS EMR | EMR Studio | Pyspark Notebook
มุมมอง 7K8 หลายเดือนก่อน
What is AWS EMR | Extract and Transform Redfin data with AWS EMR | EMR Studio | Pyspark Notebook
LeetCode 197 - Rising Temperature | Common SQL Interview Question | Self-Join | LAG window function
มุมมอง 2628 หลายเดือนก่อน
LeetCode 197 - Rising Temperature | Common SQL Interview Question | Self-Join | LAG window function
Redfin Analytics|python ETL pipeline with airflow|Data Engineering Project|Snowpipe|Snowflake|Part 2
มุมมอง 1.9K8 หลายเดือนก่อน
Redfin Analytics|python ETL pipeline with airflow|Data Engineering Project|Snowpipe|Snowflake|Part 2
Redfin Analytics|python ETL pipeline with airflow|Data Engineering Project|Snowpipe|Snowflake|Part 1
มุมมอง 5K8 หลายเดือนก่อน
Redfin Analytics|python ETL pipeline with airflow|Data Engineering Project|Snowpipe|Snowflake|Part 1
Customer Churn Data Analytics|Data Pipeline using Apache Airflow, Glue, S3, Redshift, PowerBI|Part 3
มุมมอง 1.6K9 หลายเดือนก่อน
Customer Churn Data Analytics|Data Pipeline using Apache Airflow, Glue, S3, Redshift, PowerBI|Part 3
Customer Churn Data Analytics|Data Pipeline using Apache Airflow, Glue, S3, Redshift, PowerBI|Part 2
มุมมอง 3.6K9 หลายเดือนก่อน
Customer Churn Data Analytics|Data Pipeline using Apache Airflow, Glue, S3, Redshift, PowerBI|Part 2
YouTube Channel Join Button | YouTube Memberships
มุมมอง 1179 หลายเดือนก่อน
TH-cam Channel Join Button | TH-cam Memberships
Customer Churn Data Analytics|Data Pipeline using Apache Airflow, Glue, S3, Redshift, PowerBI|Part 1
มุมมอง 6K9 หลายเดือนก่อน
Customer Churn Data Analytics|Data Pipeline using Apache Airflow, Glue, S3, Redshift, PowerBI|Part 1
Zillow Data Analytics (RapidAPI) | End-To-End Python ETL Pipeline | Data Engineering Project |Part 3
มุมมอง 3.1K9 หลายเดือนก่อน
Zillow Data Analytics (RapidAPI) | End-To-End Python ETL Pipeline | Data Engineering Project |Part 3
I saw the failed message again, but then I checked and the database had been restored. Thanks!
The scheduler does not appear to be running. Last heartbeat was received 32 minutes ago. The DAGs list may not update, and new tasks will not be scheduled.I have followed all commands you mentioned but i was getting like this in my airflow user interface and the command prompt which i connected to ec2 was running airflow standalone where to run new commands so that i can run airflow scheduler ? pls reply
I´m getting this error when execute airflow standalone ModuleNotFoundError: No module named 'tenacity.asyncio'
I solved this way pip uninstall tenacity Found existing installation: tenacity 8.4.0 pip install tenacity==8.2.1
amazing bro thanhs
After 7 hours trying to connect my instance, you really help me!
Glad it helped.
which app did you use to create the data pipeline visualization?
i'm still a bit confused, so is the AWS can be used freely, sir?🙏
Really helpful video!
Explanatory!
Very detailed!
Really helpful video!
Really helpful video!
Was a really helpful Bootcamp! 👏💯
Really helpful video!
Really helpful information!
Hi, Thanks for the video. It definitely solved my confusion points but still I have one doubt. Why airflow ? I mean we could've used the lambda for that work too right?
Excellent Tutorial
Thanks for your comment.
Thank you very much for making the concepts so easy to understand👌
You are welcome!
I am getting following error while creating a virtual environment. E: Unable to locate package python3.11.9-venv E: Couldn't find any package by glob 'python3.11.9-venv' If anyone can help then it would be great!!
Hey jay try this sudo apt install3-venv
You are superb brother..
Thank you so much 😀
Please make a video on self Join
Thank u sooo much. It helped
Glad it helped
Really great content! I just have one question, at 26:40, the part where you checked the Inbound Rules of VPC Security Group for the Redshift Cluster, is it standard practice to allow all inbound traffic and all IPv4 (Type = All traffic, Source=0.0.0.0/0)? AWS kept showing the warning to limit inbound traffic to only known IP, so I tried to set Inbound Rules to MyIP or the Public IP address of the EC2 instance I'm using to run Airflow, but for some reason step 4 failed to work when I do so (it worked fine if I follow the setting in your video for the inbound rule). Hope to hear your thoughts on this
The best ETL video I have ever come across. Thank you sir ❤🔥❤🔥❤🔥💯💯
Glad you liked it!
Absolutely wonderful tutorial, thank you for the great content man! I learned a ton from your videos and now subscribed from your channel
Thanks for the sub!
i get my load_data task failed i configured everything right but still get failed for the last task i couldnt figure it out anyone with the same scenario got any soln?
This is just what I have been searching for, thank you good sir, please kindly post more videos, you are awesome
Thank you!
te quiero mucho :D estos videos siempre son la respuesta
Thank you!
Solution to ODBC Error: Go to the security group of the Redshift cluster, click the box for the security group and edit inbound rules, Port range => 5439, source => Ipv4 Everywhere, save
Thanks alot for this video, I learnt alot from it
Glad to hear that!
Amazing content!! My friend 🙌🙌
Thank you 🙌
Excellent information, thank you so much for posting this video here
Glad it was helpful!
If you have problems with installing dependencies it is because instead of sudo apt install3.10-venv, replate it to sudo apt install3-venv to get the latest version. Currently, it's at 3.12
Invalid operation upon using that.
This worked thanks
While setting up the crawlers you said we will set up the frequency in airflow but in airflow we are not doing anything related to Glue crawler. Can you please suggest how to automate the glue crawlers ? Also When we are creating crawler and fetching data in Athena it is not showing the data correctly. You can check the timeline 1:01:07. Wrong values are populated in longitude, Gender etc. Can you please have a look ?
Thank you very much for creating this project. I followed all 3 videos from this series and learnt a lot. Thank you!
Great to hear! You r welcome.
What can you do when it say failed instead of completed?
I explained the solution to this problem in tutorial 3.1. Look for it in my playlist.
Since I’m a mech engineer coding is almost like mandarin to me but u sir the Great explanation 🙏🏻🔥🫡🫡 really loved it n totally understood ❤❤
Glad to hear that.
I got the solution for 1:05:00 , In the code Just use aws_default connection, our EC2 already has Role access to our s3 buckets. aws_conn_id='aws_default' This worked perfectly
This is how the correct logs lokks like from Airflow logs called outside TaskInstance! [2024-05-10, 19:41:17 UTC] {s3.py:110} INFO - Poking for key : s3://zillow-transformed-ben/response_data_10052024194109.csv [2024-05-10, 19:41:17 UTC] {base.py:84} INFO - Using connection ID 'aws_default' for task execution. [2024-05-10, 19:41:17 UTC] {credentials.py:1075} INFO - Found credentials from IAM Role: Ec2RoleFullAccesstoS3-2024 [2024-05-10, 19:41:22 UTC] {s3.py:110} INFO - Poking for key : s3://zillow-transformed-ben/response_data_10052024194109.csv [2024-05-10, 19:41:27 UTC] {s3.py:110} INFO - Poking for key : s3://zillow-transformed-ben/response_data_10052024194109.csv [2024-05-10, 19:41:27 UTC] {base.py:294} INFO - Success criteria met. Exiting.
Great sessions!
Thanks!!
nice demo man
Glad you enjoyed it
Hi are you working with browse job in Bangalore. The guy who is teaching here , told you will be joining their team to teach us, is this true?
yay, i managed to finish it! and i have the csv file in s3. thx, u deserve the like lol.
Nice work! You did it! Keep learning! Keep growing!
1:38:20 - i dont get why do i need aws configure INSIDE my ec2? why do i need access key when i m already inside my ec2?
Please stop saying OK.
Awesome Tutorial. Helped me a lot. Thank you..!!
Glad it helped!
Hello, you are the best Data Engineering Instructor here on TH-cam. I want to continue learning on your end to end project unfortunately I am having a problemin in initiating airflow. scheduler | [2024-05-05 04:41:04 +0000] [3014] [INFO] Booting worker with pid: 3014 scheduler | [2024-05-05T04:41:04.168+0000] {settings.py:60} INFO - Configured default timezone UTC scheduler | [2024-05-05 04:41:04 +0000] [3016] [INFO] Booting worker with pid: 3016 scheduler | [2024-05-05T04:41:04.405+0000] {manager.py:393} WARNING - Because we cannot use more than 1 thread (parsing_processes = 2) when using sqlite. So we set parallelism to 1. this is the error came from my terminal. Hope you can assist me sir. Thank you and more power
Consider Switching to a More Robust Database
dont ask for likes, i will only like if i can finish and add this to my resume
Very strict haha
Thanks man!
Happy to help!
Thanks for everything 💖 i have question if i add some data into csv they will automatically updated in power bi ?
You will need to click refresh in your powerbi. However, if you are using the powerbi service which is the paid version, you can setup an automatic refresh. For example, you can say every day 5am refresh.