If I build a Hive table on top of some S3 (or HDFS) directory like so:
create external table newtable (name string)
row format delimited
fields terminated by ','
stored as textfile location 's3a://location/subdir/';
When I add files to that S3 location, the Hive table doesn't automatically update. The new data is only included if I create a new Hive table on that location. Is there a way to build a Hive table (maybe using partitions) so that whenever new files are added to the underlying directory, the Hive table automatically shows that data (without having to recreate the Hive table)?
s3a://location/subdir/or to any subdirectories under this location? - franklinsijos3a://location/subdir/. @Dudu Every file is supposed to be scanned which is why if I add another file to that subdirectory, I would expect that data to show up when I run 'select *' on the table. But it doesn't; it shows the same table (without the newly added data). - covfefe