• Home
  • Blog
  • AWS
  • Contact
Brandorr Group
  • Home
  • Blog
  • AWS
  • Contact

AWS Wrangler for FinOps: Optimize AWS Costs & Data Processing with Python

2/25/2025

0 Comments

 
AWS Wrangler (now officially AWS SDK for pandas) is an open-source Python library that simplifies working with AWS services like S3, Glue, Athena, Redshift, DynamoDB, and more. While it integrates seamlessly with pandas, it also supports other Python data libraries such as Apache Arrow, Modin, Ray, and PySpark, making it a versatile tool for AWS data processing.

AWS Wrangler for FinOps and Cost Optimization
FinOps and AWS cost optimization require working with vast amounts of data, including price lists, consumption data, inventory data, and organization data. Python is an excellent language for interacting with AWS data services, and AWS Wrangler makes this process even more efficient.

While there are edge cases where boto3 is still useful, in most cases, AWS Wrangler greatly simplifies and shortens your code—especially when incorporating Athena queries into Python workflows.
Picture

Key Features of AWS Wrangler
  • Effortless Read & Write to S3 – Supports Parquet, CSV, JSON, and integrates with Glue Catalog.
  • Optimized Queries on Athena – Pushes down filters to minimize scanned data and reduces Athena costs using query caching with CTAS (Create Table As Select).
  • Leverage Glue Catalog – Streamlines schema management for better efficiency and transparency.
  • Simplify DataFrame Operations – Read from and write to Parquet, CSV, or JSON files in S3 effortlessly.
  • Works with Multiple AWS Data Services – Compatible with S3, Glue, Athena, Redshift, Timestream, OpenSearch, Neptune, QuickSight, CloudWatch Logs, DynamoDB, EMR, Secrets Manager, and RDS (Aurora, PostgreSQL, MySQL, SQL Server, Oracle).
The result? Optimized storage, faster queries, and lower costs.

​Why AWS Wrangler is Great for Cost Optimization
  • Reduces Data Transfer Costs – Enables in-place filtering and processing in S3, cutting unnecessary data movement.
  • Lowers Athena Query Costs – Minimizes scanned data, directly reducing query expenses.
  • Saves Analyst Time – Simplifies data coding tasks, allowing teams to focus on higher-value work.

Pro Tip for Maximum Cost Efficiency
Combine AWS Wrangler with tools like:
  • Boto3
  • Cost Explorer API
  • S3 Inventory & S3 Lens Metrics exports
  • Glue Data Catalog
  • AWS Cost Optimization Hub
Using these together helps you analyze cost data and uncover optimization opportunities in real time. If you're already using pandas or other DataFrame libraries, AWS Wrangler feels like a natural extension—and if you’re managing AWS resources, it can make your workflow significantly more efficient.

Want to Optimize AWS Costs with AWS Wrangler?
Have you used AWS Wrangler in your cost optimization efforts? Whether you're just exploring its capabilities or actively using it, join the conversation over on LinkedIn, or connect with me directly on my LinkedIn page to discuss how it can streamline your workflows and reduce costs. Connect with us to explore AWS cost optimization strategies tailored to your needs!
0 Comments

Netflix’s DIY AWS Cost Optimization Approach Still Resonates 7 Years Later

2/11/2025

0 Comments

 
In 2017, Netflix gave a talk at AWS re:Invent that left a lasting impression on me. The session, "Tooling Up for Efficiency: DIY Solutions @ Netflix", presented by Andrew Park and Sébastien Delarquier, shared how Netflix built in-house solutions to optimize cloud costs at scale.

This talk profoundly shaped my perspective on AWS cost management. Netflix laid out their approach to cloud management, with a particular focus on cost efficiency. They also made a compelling case for why large AWS users might benefit from developing their own cost optimization tools.

Trade-offs in Cloud Management
One key insight that I believe shapes any organization’s cost optimization efforts is understanding the trade-offs required when setting cloud management priorities:
  • Innovation – Experimenting, iterating quickly, and developing new features.
  • Reliability – Ensuring uptime, system resilience, and handling failure gracefully.
  • Security – Protecting data, preventing breaches, and meeting compliance needs.
  • Efficiency – Optimizing cloud costs and resource utilization.
No organization can maximize all four at once. If you’re going to prioritize something, something else must be deprioritized. Conversely, if everything is a priority, nothing truly is.
With this in mind, Netflix walked through how they manage cost efficiency and the tooling they developed to support it.

The Key Takeaway: Data-Driven Decision Making
Whether you’re using commercial tooling, cloud-native cost tools, DIY solutions, or a mix of all three, their approach still makes sense.
The biggest takeaway for me? Data-driven decision-making is essential for cloud cost optimization. The approach will feel familiar to many enterprise data teams—it’s a blend of business intelligence (BI), data analysis, data engineering, and data science.

A Personal Observation
Many of the capabilities Netflix outlined in their 2017 talk have since been incorporated into AWS’s native cost and billing tools, as well as the Well-Architected Framework. These tools are a great starting point—and for most shops, they get you most of the way there.

However, I’ve found that developing additional tooling for deeper data analysis, automation, and alerting can significantly enhance AWS’s built-in capabilities.

What’s Your Cloud Cost Management Strategy?
How does your organization approach AWS cost management? Do you prefer native AWS tools, commercial dashboards, DIY solutions, or a combination? Join the conversation over on LinkedIn, or connect with me directly on my LinkedIn page to share your experience or explore how we can help build a tailored cost optimization strategy for your organization.
0 Comments

Run Kubernetes at Scale with Spot Instances: Karpenter 1.X is Here!

2/4/2025

0 Comments

 
After three years of public testing, Karpenter 1.0 was officially released in August, and since then, it has been rapidly evolving. Now at version 1.2.1, this open-source cluster autoscaler is ready for production workloads, providing a smarter way to manage Kubernetes infrastructure by dynamically selecting the most cost-efficient instance types without sacrificing performance.

Key Features of Karpenter 1.X
  • Dynamic provisioning – Automatically provisions the right compute resources for your application in real time.
  • Optimized capacity management – Reduces over-provisioning and idle capacity, lowering cloud costs.
  • Seamless scaling – Native Kubernetes integration simplifies scaling strategies.
  • Enhanced disruption controls – Users can now set disruption budgets by reason, enabling finer control over node disruptions for underutilization, emptiness, or drift scenarios.
  • Improved drift management – Nodes that deviate from a desired state are now automatically replaced by default, ensuring consistency and reliability.
​Karpenter is built for modern cloud-native environments, where flexibility is key. Whether you're scaling your infrastructure to handle peak traffic or optimizing resources during quieter periods, Karpenter helps you maximize efficiency and reduce costs.

Ready to Optimize Kubernetes at Scale? 
​
Are you using Karpenter in your Kubernetes environment, or are you considering making the switch? Let’s explore how Karpenter can help optimize costs and enhance scalability for your AWS infrastructure. Join the conversation over on LinkedIn, or connect with me directly on my LinkedIn page to discuss best practices and implementation strategies for cost-effective Kubernetes management!
0 Comments

      Subscribe

      Only good things, and we'll never sell your email, we promise.
    Stay Updated
    View my profile on LinkedIn

    Author

    Brandorr Group LLC is a one-stop cloud computing solution provider, helping companies manage growth and ship new projects using cloud and scalability best practices.

    Recent Posts

    April 2025
    March 2025
    February 2025
    January 2025
    December 2024
    November 2024
    October 2024
    June 2019
    May 2019
    April 2017
    January 2017
    December 2016
    August 2016
    December 2015
    September 2015
    March 2015
    January 2015
    January 2012
    October 2011
    April 2011
    December 2010
    July 2010
    March 2010
    March 2009

Location

3117 Broadway 
Suite 50
New York, NY ​10027

Contact Us

​We'd love to hear from you,
​please feel free to call us:
866.991.4838

About Brandorr

With decades of experience in cloud technologies and specialties in high volume/throughput, high availability, and disaster mitigation engineering, Brandorr Group has the experience to help customers of all sizes develop, deploy and manage their new or existing infrastructure in the cloud.

By using provisioning and configuration management technologies such as Docker, Ansible, Chef, Puppet, Terraform, and CloudFormation, we are able to quickly and cost-effectively scale and deploy infrastructure projects of any size.
​
Additionally, Brandorr maintains 24x7 systems engineering, security and monitoring teams augmented by database administrators and software developers to ensure projects are delivered and systems remain highly available while maintaining performance.
  • Home
  • Blog
  • AWS
  • Contact