▲ | JustExAWS 3 days ago | |||||||||||||||||||||||||||||||||||||||||||||||||
I am also a former AWS employee. What non public information did you need for DDB? | ||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | tracker1 3 days ago | parent | next [-] | |||||||||||||||||||||||||||||||||||||||||||||||||
Try ingesting the a complete WHOIS dump into DDB sometime. This was before autoscaling worked at all when I tried... but it absolutely wasn't anything one can consider fun. In the end, after multiple implementations, finally had to use a Java Spring app on a server with a LOT of ram just to buffer the CSV reads without blowing up on the pushback from DDB. I think the company spent over $20k in the couple months on different efforts in a couple different languages (C#/.Net, Node.js, Java) across a couple different routes (multiple queues, lambda, etc) just to get the initial data ingestion working a first time. The Node.js implementation was fastest, but would always blow up a few days in without the ability to catch with a debugger attached. The queues and lambda experiments had throttling issues similar to the DynamoDB ingestion itself, even with the knobs turned all the way up. I don't recall what the issue with the .Net implementation was at the time, but it blew up differently. I don't recall all the details, and tbh I shouldn't care, but it would have been nice if there was some extra guidance of trying to take in a few gb of csv into DynamoDB at the time. To this day, I still hate ETL work. | ||||||||||||||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | cyberax 3 days ago | parent | prev [-] | |||||||||||||||||||||||||||||||||||||||||||||||||
A tool to look at hot partitions, for one thing. | ||||||||||||||||||||||||||||||||||||||||||||||||||
|