Thanks for this, One thing to add here is you should have S3 VPC gateway endpoint attached to the VPC, and relevant route table (whether public or private) as well for this to work. Secondly at 17:40 you should select security group, not the subnet group.
Hi, appreciate your efforts on this video. I have few doubts. First one is all these RDS, S3 and Athena should be in the same AWS account or can be in different accounts? I our environment we are using RDS and Athena in different accounts so cannot use the same SG. Please reply
Hi, I created aa glue crawler to get all the data from a particular sub folder of a S3 bucket. IAM role : Glue services and also added the S3 read only acess. I am able to create the table in the DB. But these table have ZERO records. I checked I have the data in the excel, csv and parquet (all format) but still not able to get the data just the columns names and 0 records. please help me to resolve this issue
thanks for the reply if am working in pycharm and the transformed data in s3 bucket or in some data base ,how to extract that data to pycharm if the data is too big to download into my local machine
Thanks for this,
One thing to add here is you should have S3 VPC gateway endpoint attached to the VPC, and relevant route table (whether public or private) as well for this to work.
Secondly at 17:40 you should select security group, not the subnet group.
Thank you for watching my videos.
Appreciate you attention and thank you for sharing the updates here.
Hi, appreciate your efforts on this video. I have few doubts. First one is all these RDS, S3 and Athena should be in the same AWS account or can be in different accounts? I our environment we are using RDS and Athena in different accounts so cannot use the same SG. Please reply
Thank you for watching my videos.
Yes, Indeed it can be cross account.
Hi, I created aa glue crawler to get all the data from a particular sub folder of a S3 bucket. IAM role : Glue services and also added the S3 read only acess. I am able to create the table in the DB. But these table have ZERO records. I checked I have the data in the excel, csv and parquet (all format) but still not able to get the data just the columns names and 0 records.
please help me to resolve this issue
Thank you for watching my videos.
Please check the pipeline logs potentially column datatype mismatch or could be permission issue.
hey hi I have few doubts regarding data scientist could you please help me??
Thank you for watching my videos.
I shall try my best , please let me know your questions.
thanks for the reply if am working in pycharm and the transformed data in s3 bucket or in some data base ,how to extract that data to pycharm if the data is too big to download into my local machine