Millions of messages are legacy, and in the new modern world of data, we like “billions”. This is exactly the terminology in the use case we faced from a very prominent client in Egypt. The scenario demanded more attention as this valuable client did multiple proof of the concepts with many other open sources and could not meet exact SLA and needs. The client wanted to have more than a hundred billion( yes, “b”) messages in eight hours to be ingested and further queried without much latency. The presentation will be a live demonstration of how we can architect such a solution with PrestoDB under the hood and some simple but advanced ingestion capabilities and data formats.