Let's assume that I have a table with following attributes:
- UNIQUE
user_id
(primary hash key) category_id
(GSI hash index)timestamp
I will have a lot of users, but only few categories.
user_id | category_id
1 1
3 1
4 1
5 3
.. ..
50000000 1
Is it ok to store millions of records with the same category_id
value as a Global Secondary Index? Should I expect any restrictions?
I'm wondering if scan is not a bad choice. I will use filtering by category_id
only once a day. What is the cost (time and money) of scanning millions of records?
Thanks!