Announcing Bacalhau v1.8.0: Intelligent Edge Computing Meets Enterprise Integration
Discover how Bacalhau v1.8.0 transforms distributed computing with a native Splunk integration, name-based job management, and enhanced daemon orchestration.
We're excited to announce Bacalhau v1.8.0: a groundbreaking release that transforms how you can approach distributed computing at the edge. This new release introduces:
Intelligent cost optimization through an advanced Splunk integration.
Enhanced daemon job orchestration for dynamic infrastructure.
Enterprise-grade improvements—docs, security, and more.
A completely reimagined job lifecycle management system.
In a nutshell, Bacalhau v1.8.0 delivers the integrations and capabilities that you need to reduce costs, improve operational efficiency, and unlock the full potential of your distributed infrastructure.
Let’s dive into it to learn more!
Slash Your Splunk Bill: Intelligent Logging at the Edge
Let's talk about the elephant in the room: logging costs. For too long, you've faced a terrible choice: ship all your raw data to Splunk and watch your budget evaporate, or sample your data and risk missing the one critical event you need.
So, what if there could be an alternative solution? Bacalhau’s new Splunk integration offers a third, much smarter path. Instead of moving mountains of data, Bacalhau lets you ship the computation to the data source. This fundamentally changes the cost equation.
With v1.8.0, Bacalhau's compute power can be integrated directly into Splunk so that you can:
Process data at the source: Deploy smart logging agents that filter, aggregate, and analyze data on the edge, sending only high-signal results to Splunk.
Fork your data stream: Send critical, processed alerts to Splunk for real-time monitoring while simultaneously archiving the full, compressed raw logs to cheap object storage.
Query live data without ingestion: Run queries directly on remote nodes to investigate issues in real-time, without moving the data or paying ingestion fees.
Replay history on demand: Need to analyze old logs? No problem. Use Bacalhau's compute power to pull data from your archive, process it, and analyze it in Splunk.
This isn't a minor tweak. It's a paradigm shift that can cut logging costs by 60-80% while giving you more analytical power. The best part? It all feels native. You use the Splunk dashboards and search queries you already know, while Bacalhau’s distributed compute engine works transparently in the background.
Enhanced Daemon Jobs
Bacalhau v1.8.0 strengthens daemon job reliability with enhanced orchestration. The goal is to ensure comprehensive coverage across dynamic infrastructure.
Our developers have beefed up the entire system for the reliability that production environments demand:
Node discovery is faster.
Constraint checking is more robust.
Deployment tracking gives you a clear view of your job's footprint. When a new node spins up in your dynamic infrastructure, you can be confident your daemon job will deploy to it automatically and reliably.
These enhancements are particularly valuable for the Splunk integration, where logging agents must reliably deploy to new infrastructure automatically.
Enterprise-Ready by Default
Rounding out the release are features designed to make Bacalhau seamless in an enterprise environment:
Managed result storage: With Expanso Cloud, you no longer need to configure your own publisher or manage storage credentials. Job results are automatically and securely stored in our managed infrastructure. Just run your job and fetch your results.
Documentation overhaul: We’ve rebuilt our docs from the ground up with task-based guides, architectural deep dives, and real-world examples for things like distributed log processing and edge ML.
Hardened foundation: We've continued to strengthen security, authentication (with SSO), and reliability to ensure Bacalhau is ready for your most demanding production workloads.
Transformative Job Lifecycle Management
Bacalhau v1.8.0 introduces a fundamental transformation in how you manage computational workflows. We've moved beyond cryptic, auto-generated job IDs to a powerful name-based system that supports versioning, updates, and intelligent rerunning.
Goodbye j-f47ac10b-58cc-4372-a567-0e02b2c3d479
If you've ever felt the pain of managing workflows using nothing but UUIDs, this update is for you. We've completely reimagined job lifecycle management to be intuitive, version-controlled, and built for humans.
Jobs are now identified by meaningful names you define in your spec:
# Before: Cryptic UUID management
bacalhau job describe j-f47ac10b-58cc-4372-a567-0e02b2c3d479
# After: Intuitive name-based operations
bacalhau job describe monthly-sales-report-generator
Advanced Versioning and Dry-Run
Every job submission with an existing name creates a new version. The enhanced --dry-run functionality provides server-side diff previews, showing exactly what will change before applying updates:
# Preview changes with detailed diff
bacalhau job run --dry-run updated-pipeline.yaml
# Work with specific versions
bacalhau job logs data-analysis --version 2
bacalhau job describe ml-model-training --version 1
Intelligent Job Rerunning
The new rerun command eliminates the need to resubmit jobs. This avoids you from creating new versions of existing jobs while preserving history:
# Rerun the latest version
bacalhau job rerun data-processing-pipeline
# Rerun a specific version
bacalhau job rerun ml-training-job --version 3
Also, this versioning system is fully backward compatible: existing workflows using job IDs continue to work while new capabilities are available for adoption at your own pace.
What’s next?
Bacalhau v1.8.0 is available now with full backward compatibility. Existing workflows continue to operate unchanged while new capabilities are available for immediate adoption:
For Splunk Users: The Expanso Splunk application is available through Splunkbase and can be deployed in minutes to start realizing immediate cost savings.
For Existing Users: Update to v1.8.0 to access enhanced daemon job orchestration, managed result storage, and the new job lifecycle management capabilities.
For New Users: Our completely rebuilt documentation provides clear pathways for getting started with distributed computing, whether for cost optimization, edge processing, or advanced analytics.
Join Us on the Journey: 5 Days of Bacalhau
Stay tuned for our "5 Days of Bacalhau" series, in which we'll go deeper into these exciting new features:
Day 1: Announcing Bacalhau 1.8.0—Intelligent Edge Computing Meets Enterprise Integration (this post)
Day 2: Rerun, Update, and Version Your Bacalhau Jobs
Day 3: How Bacalhau Boosts Daemon Job Reliability
Day 4: Seamless Result Storage with Managed Publishers
Day 5: Distributed Logs Management with Bacalhau and Splunk
Get Involved!
We welcome your involvement in Bacalhau. There are many ways to contribute, and we’d love to hear from you. Reach out at any of the following locations:
Commercial Support
While Bacalhau is open-source software, the Bacalhau binaries go through the security, verification, and signing build process lovingly crafted by Expanso. Read more about the difference between open-source Bacalhau and commercially supported Bacalhau in the FAQ. If you want to use the pre-built binaries and receive commercial support, contact us or get your license on Expanso Cloud!