Loading...


Updated 1 March 2026 • 5 mins read
Khushi Dubey | Author
Table of Content

At Opslyft, we work closely with growing SaaS businesses that rely heavily on AWS storage. In this case, the customer operated a global video hosting platform for enterprise sales and marketing teams. With nearly one million full HD videos, their storage footprint approached 10 petabytes.
Although Amazon S3 provided the scalability and reliability they needed, costs were rising steadily. Our goal was simple: understand the architecture, identify cost drivers, and redesign the system so that the storage strategy matched real access behavior. What followed was a structured optimization process that reduced their Amazon S3 bill by approximately 70 percent.
We enabled Amazon S3 Access Logging to capture detailed records of every request made to the storage buckets. These logs were then analyzed using Amazon Athena, allowing us to query access frequency at scale.
Within hours, the data revealed a clear pattern. A very small percentage of video files, roughly 0.1 percent, accounted for nearly half of all GET and retrieval activity in Glacier Instant Retrieval. These were typically larger marketing videos in high resolution, such as 1080p or 4K.
Although they represented only a small fraction of the total 10 million objects, they generated a disproportionate number of requests. In fact, around 10 percent of objects were responsible for approximately 99 percent of the 3.1 billion monthly GET requests.
Storing these high-traffic files in Glacier Instant Retrieval was inefficient. We evaluated moving them to S3 Intelligent-Tiering, which does not charge retrieval fees and has significantly lower GET request costs. Cost modeling showed potential savings of over 90 percent for the most accessed files.
We flagged the top 60,000 active objects, about 0.6 percent of total content, and moved them to S3 Intelligent-Tiering. This change alone reduced retrieval and GET costs by roughly 50 percent.
The deeper insight was that not all videos behave the same way. Some content remains rarely viewed, while others attract sustained traffic over time. A single storage strategy could not serve both patterns efficiently.
We restructured the model as follows:
This required only minor code adjustments but produced a significant financial impact. GET request volume gradually shifted away from Glacier Instant Retrieval, and overall S3 costs began to decline steadily.
After optimizing storage classes, we focused on another major cost driver: the sheer number of S3 GET requests. Even with proper storage placement, excessive request volume increases expenses.
We examined two layers of the system:
The platform used Amazon CloudFront as its content delivery network. Analysis of Amazon CloudWatch metrics showed a global cache hit rate of only 65 percent, with some regions dropping to 40 percent.
A low cache hit rate means more requests fall back to the origin, which in this case was Amazon S3. After tuning the CloudFront distribution settings and improving regional configurations, the cache hit rate increased to approximately 90 percent.
This improvement alone reduced S3 GET and retrieval requests by about 50 percent.
The Just-In-Time packaging layer, based on Nginx, regenerated video segments whenever CloudFront experienced a cache miss. Because it did not cache files locally, each segment required multiple S3 GET range requests.
Originally, each segment triggered an average of 7.05 GET requests. By increasing the byte range size from 256 KB to 2 MB, we reduced that average to 1.04 GET requests per segment.
This change reduced GET request volume by approximately 85 percent at the packaging layer.
Combined with storage optimization and CDN tuning, total S3 GET requests dropped by around 90 percent.
Through detailed cost analysis using AWS Cost and Usage Reports and S3 Access Logging, combined with architectural adjustments across storage and delivery layers, the platform reduced its six-figure annual Amazon S3 bill by roughly 70 percent.
The most important lesson from this engagement is clear. Cost optimization on AWS is rarely about removing services. It is about aligning architecture with real usage patterns.
In large-scale environments, small percentages matter. A fraction of objects can drive the majority of the cost. Without detailed visibility, those cost drivers remain hidden.
At Opslyft, we approach optimization through architecture review, data-driven analysis, and targeted adjustments. When storage classes, access patterns, and content delivery layers are aligned correctly, organizations can scale confidently while maintaining financial control.