#change-event-streaming

[ follow ]
DevOps
fromInfoQ
1 day ago

Beyond One-Click: Designing an Enterprise-Grade Observability Extension for Docker

Docker Extensions enhance developer productivity but may not meet enterprise needs for security, compliance, and integration.
Startup companies
fromInfoQ
1 day ago

Platform Engineering: Lessons from the Rise and Fall of eBay Velocity

eBay pioneered many technologies but ultimately could not save the company despite doubling engineering productivity.
Software development
fromMedium
2 days ago

Async Logging Is Not a Silver Bullet - What Actually Limits Performance

Async logging redistributes costs rather than reducing them, impacting performance in different ways depending on implementation.
Business intelligence
fromInfoWorld
16 hours ago

The hyperscalers are pricing themselves out of AI workloads

AI is challenging traditional cloud pricing models, as buyers seek exceptional value beyond brand recognition and familiar pricing strategies.
fromInfoQ
5 days ago

Latency: The Race to Zero...Are We There Yet?

In the fintech industry we can link latency directly to profit and money. If I have lower latency than the competition, I can get to the better deals, I can make the better deals.
Venture
fromInfoQ
1 day ago

Airbnb Migrates High-Volume Metrics Pipeline to OpenTelemetry

The resulting system now ingests over 100 million samples per second in production, showcasing the scalability and efficiency of the new metrics stack.
DevOps
fromInfoWorld
2 weeks ago

How Apache Kafka flexed to support queues

Apache Kafka has cemented itself as the de facto platform for event streaming, often referred to as the 'universal data substrate' due to its extensive ecosystem that enables connectivity and processing capabilities.
Scala
Business intelligence
fromDevOps.com
4 days ago

Dynatrace to Acquire Bindplane to Process and Route Telemetry Data - DevOps.com

Dynatrace is acquiring Bindplane to enhance log management and allow pre-processing of telemetry data for cost reduction and workflow efficiency.
Design
fromInfoQ
1 week ago

Panel: Taking Architecture Out of the Echo Chamber

Architecture's importance is growing, necessitating a shift in practice to avoid past mistakes and engage with broader conversations.
#aws
fromTheregister
5 days ago
DevOps

AWS put a file system on S3; I stress-tested it

AWS S3 Files allows mounting S3 buckets as NFS shares, providing solid conflict resolution and cost-effective storage options.
DevOps
fromInfoWorld
2 weeks ago

What front-end engineers need to know about AWS

Understanding AWS infrastructure improves front-end debugging and UI performance.
DevOps
fromAmazon Web Services
1 day ago

Troubleshooting environment with AI analysis in AWS Elastic Beanstalk | Amazon Web Services

AWS Elastic Beanstalk simplifies web application deployment and scaling, now enhanced with AI Analysis for troubleshooting environment health issues.
DevOps
fromTechzine Global
4 days ago

AWS launches Agent Registry for managing AI agents

AWS introduces the Agent Registry to centralize AI agent management and reduce chaos in organizations deploying numerous agents.
DevOps
fromInfoWorld
4 days ago

AWS targets AI agent sprawl with new Bedrock Agent Registry

AWS introduces Agent Registry to help enterprises manage and govern AI agents effectively.
DevOps
fromTheregister
5 days ago

AWS put a file system on S3; I stress-tested it

AWS S3 Files allows mounting S3 buckets as NFS shares, providing solid conflict resolution and cost-effective storage options.
DevOps
fromInfoWorld
2 weeks ago

What front-end engineers need to know about AWS

Understanding AWS infrastructure improves front-end debugging and UI performance.
Software development
fromInfoQ
6 days ago

Stateful Continuation for AI Agents: Why Transport Layers Now Matter

Transport layer efficiency is crucial for agent workflows, as multi-turn interactions significantly increase overhead compared to single-turn LLM use.
Business intelligence
fromZDNET
5 days ago

I asked 5 data leaders about how they use AI to automate - and end integration nightmares

Strong processes and AI integration are essential for businesses to effectively utilize data.
Scala
fromInfoQ
1 week ago

Beyond RAG: Architecting Context-Aware AI Systems with Spring Boot

Context-Augmented Generation (CAG) enhances Retrieval-Augmented Generation (RAG) by managing runtime context for enterprise applications without requiring model retraining.
#kubernetes
DevOps
fromMedium
2 days ago

KubeCraft: Talk to Your Kubernetes Cluster Like a Colleague

KubeCraft simplifies Kubernetes management by allowing users to interact with their clusters using plain English through an AI assistant.
DevOps
fromInfoWorld
5 days ago

Bringing databases and Kubernetes together

Automating Kubernetes workloads with Operators can provide DBaaS functionality while avoiding provider lock-in.
DevOps
fromMedium
2 days ago

KubeCraft: Talk to Your Kubernetes Cluster Like a Colleague

KubeCraft simplifies Kubernetes management by allowing users to interact with their clusters using plain English through an AI assistant.
DevOps
fromInfoWorld
5 days ago

Bringing databases and Kubernetes together

Automating Kubernetes workloads with Operators can provide DBaaS functionality while avoiding provider lock-in.
#multi-agent-systems
Information security
fromTechzine Global
3 weeks ago

Databricks launches Lakewatch: agentic SIEM on the Lakehouse

Lakewatch is an open SIEM platform that consolidates security, IT, and business data, enabling rapid threat detection and response using AI agents.
#apache-spark
Java
fromMedium
3 weeks ago

Spark Internals: Understanding Tungsten (Part 1)

Apache Spark revolutionized big data processing but faces challenges due to JVM memory management and garbage collection issues.
Java
fromMedium
3 weeks ago

Spark Internals: Understanding Tungsten (Part 2)

Catalyst Optimizer and Tungsten work together in Apache Spark to optimize data execution and manage raw binary data.
Java
fromMedium
3 weeks ago

Spark Internals: Understanding Tungsten (Part 1)

Apache Spark revolutionized big data processing but faces challenges due to JVM memory management and garbage collection issues.
Java
fromMedium
3 weeks ago

Spark Internals: Understanding Tungsten (Part 2)

Catalyst Optimizer and Tungsten work together in Apache Spark to optimize data execution and manage raw binary data.
DevOps
fromDevOps.com
4 days ago

CloudBees Delivers on AI Promise to Improve Application Testing - DevOps.com

CloudBees Smart Tests uses AI to prioritize tests, reducing CI/CD processing time significantly.
Data science
fromMedium
1 month ago

Migrating to the Lakehouse Without the Big Bang: An Incremental Approach

Query federation enables safe, incremental lakehouse migration by allowing simultaneous queries across legacy warehouses and new lakehouse systems without risky big bang cutover approaches.
fromInfoWorld
3 weeks ago

We mistook event handling for architecture

Events are essential inputs to modern front-end systems. But when we mistake reactions for architecture, complexity quietly multiplies. Over time, many front-end architectures have come to resemble chains of reactions rather than models of structure. The result is systems that are expressive, but increasingly difficult to reason about.
React
Data science
fromMedium
4 weeks ago

Building Consistent Data Foundations at Scale

Building consistent data foundations through intentional architecture, engineering, and governance is essential to prevent fragmentation, support AI adoption, ensure regulatory compliance, and enable reliable organizational decisions at scale.
Scala
fromInfoQ
3 weeks ago

QCon London 2026: Introducing Tansu.io -- Rethinking Kafka for Lean Operations

Tansu is an open-source, stateless messaging broker that replaces Kafka's complex architecture with a simpler, durable storage model.
DevOps
fromInfoQ
6 days ago

AAIF's MCP Dev Summit: Gateways, gRPC, and Observability Signal Protocol Hardening

MCP Dev Summit 2026 showcased the protocol's readiness for enterprise-scale production with significant advancements and commitments from major companies like Amazon.
DevOps
fromInfoWorld
6 days ago

AWS turns its S3 storage service into a file system for AI agents

S3 Files simplifies access to Amazon S3, enhancing its role as a primary data layer for AI and modern applications.
DevOps
fromInfoQ
6 days ago

Uber's Hive Federation Decentralizes 16K Datasets and 10+ PB for Zero-Downtime Analytics at Scale

Uber redesigned its Hive data warehouse to decentralize datasets, enhancing scalability, security, and operational autonomy for teams.
Scala
fromMedium
4 weeks ago

What I Learned Building Secure Observability in Scala

Build secure Scala applications by keeping core logic in plain IO and using a temporary Mission Stack only for sensitive operations, integrating security with observability from the start rather than adding it later.
Artificial intelligence
fromComputerWeekly.com
1 month ago

Edge AI: What's working and what isn't | Computer Weekly

Edge AI deployment success depends on identifying efficient, narrow use cases with manageable risks rather than pursuing sophisticated, large-scale models across all applications.
Business intelligence
fromInfoWorld
3 weeks ago

Snowflake's new 'autonomous' AI layer aims to do the work, not just answer questions

Project SnowWork is Snowflake's autonomous AI layer that automates data analysis tasks like forecasting, churn analysis, and report generation without requiring data team intervention.
fromTechzine Global
6 days ago

AWS S3 buckets now support file systems

S3 Files is built on Amazon EFS and automatically translates file system operations into S3 requests, allowing applications to work with S3 data without code changes.
DevOps
Data science
fromMedium
1 month ago

100 Scala Interview Questions and Answers for Data Engineers

Structured Scala and Apache Spark interview preparation requires understanding distributed systems, performance trade-offs, and pipeline design beyond theoretical knowledge.
DevOps
fromDevOps.com
1 week ago

Apica Extends Scope and Reach of Platform for Managing Telemetry Data - DevOps.com

Apica's Ascent platform update enhances telemetry data management for DevOps teams, improving observability and cost control.
fromInfoQ
1 month ago

Hybrid Cloud Data at Uber: How Engineers Solved Extreme-Scale Replication Challenges

Uber's engineering team has transformed its data replication platform to move petabytes of data daily across hybrid cloud and on-premise data lakes, addressing scaling challenges caused by rapidly growing workloads. Built on Hadoop's open-source Distcp framework, the platform now handles over one petabyte of daily replication and hundreds of thousands of jobs with improved speed, reliability, and observability.
Miscellaneous
Software development
fromMedium
1 month ago

Unified Databricks Repository for Scala and Python Data Pipelines

Databricks repositories require structured setup with Gradle for multi-language support, dependency management, and version control to scale beyond manual notebook maintenance.
Artificial intelligence
fromInfoWorld
1 month ago

Why AI requires rethinking the storage-compute divide

AI workloads require continuous processing of unstructured multimodal data, causing redundant data movement and transformation that wastes infrastructure costs and data scientist time.
DevOps
fromInfoQ
1 week ago

Replacing Database Sequences at Scale Without Breaking 100+ Services

Validating requirements can simplify complex problems, and embedding sequence generation reduces network calls, enhancing performance and reliability.
DevOps
fromTechzine Global
1 week ago

Observability warehouses, the next structural evolution for telemetry

Observability is essential for real-time insights in cloud systems, helping to reduce downtime and improve performance.
#event-driven-architecture
DevOps
fromInfoQ
2 weeks ago

Event-Driven Patterns for Cloud-Native Banking: Lessons from What Works and What Hurts

Event-driven architecture introduces complexity and requires careful implementation, especially in regulated environments, to ensure reliability and system evolution.
fromInfoQ
1 month ago
Software development

[Video Podcast] Building Resilient Event-Driven Microservices in Financial Systems with Muzeeb Mohammad

Event-driven architectures using Kafka enable decoupling backend workflows, improving scalability and SLAs for complex multi-system processes like account opening.
DevOps
fromInfoQ
2 weeks ago

Event-Driven Patterns for Cloud-Native Banking: Lessons from What Works and What Hurts

Event-driven architecture introduces complexity and requires careful implementation, especially in regulated environments, to ensure reliability and system evolution.
fromInfoQ
1 month ago
Software development

[Video Podcast] Building Resilient Event-Driven Microservices in Financial Systems with Muzeeb Mohammad

Web frameworks
fromLoicpoullain
1 month ago

The future of web frameworks in the age of AI

AI agents now generate 90-95% of production code, requiring frameworks to be AI-understandable with comprehensive documentation and clear examples to remain competitive.
fromInfoQ
2 months ago

350PB, Millions of Events, One System: Inside Uber's Cross-Region Data Lake and Disaster Recovery

Uber has built HiveSync, a sharded batch replication system that keeps Hive and HDFS data synchronized across multiple regions, handling millions of Hive events daily. HiveSync ensures cross-region data consistency, enables Uber's disaster recovery strategy, and eliminates inefficiency caused by the secondary region sitting idle, which previously incurred hardware costs equal to the primary, while still maintaining high availability. Built initially on the open-source Airbnb ReAir project, HiveSync has been extended with sharding, DAG-based orchestration, and a separation of control and data planes.
Tech industry
DevOps
fromInfoQ
3 weeks ago

Uber Launches IngestionNext: Streaming-First Data Lake Cuts Latency and Compute by 25%

Uber's IngestionNext platform shifts to a streaming-first system, reducing data ingestion latency from hours to minutes for analytics and machine learning.
Data science
fromDevOps.com
2 months ago

Why Data Contracts Need Apache Kafka and Apache Flink - DevOps.com

Data contracts formalize schemas, types, and quality constraints through early producer-consumer collaboration to prevent pipeline failures and reduce operational downtime.
Startup companies
fromInfoQ
2 months ago

Etleap Launches Iceberg Pipeline Platform to Simplify Enterprise Adoption of Apache Iceberg

Managed Iceberg pipeline platform unifies ingestion, transformation, orchestration, and table operations inside customers' VPCs, enabling enterprise Iceberg adoption without building custom stacks.
Artificial intelligence
fromInfoQ
2 months ago

Autonomous Big Data Optimization: Multi-Agent Reinforcement Learning to Achieve Self-Tuning Apache Spark

A Q-learning agent autonomously learns and generalizes optimal Spark configurations by discretizing dataset features and combining with Adaptive Query Execution for superior performance.
DevOps
fromInfoQ
3 weeks ago

AWS Expands Aurora DSQL with Playground, New Tool Integrations, and Driver Connectors

Amazon Aurora DSQL introduces usability enhancements, including a browser-based playground and integrations with popular SQL tools for improved developer experience.
Software development
fromInfoWorld
2 months ago

Why your next microservices should be streaming SQL-driven

Streaming SQL with UDFs, materialized results, and ML/AI integrations enables continuous, stateful processing of event streams for microservices.
DevOps
fromMedium
4 weeks ago

Reusing CloudFront, ALB, and API Gateway in a Serverless Platform

Edge architecture decisions around CloudFront, ALB, and API Gateway significantly impact deployment speed and isolation in serverless platforms while balancing security, compliance, and cost.
DevOps
fromInfoQ
4 weeks ago

QCon London 2026: Uncorking Queueing Bottlenecks with OpenTelemetry

Distributed tracing with OpenTelemetry enables engineers to identify root causes across service boundaries by maintaining hierarchical visibility of operations, while SLOs based on latency provide more reliable alerting than infrastructure metrics.
Software development
fromMedium
2 months ago

When Kafka Lag Lies: A Production Debugging Story

Uncommitted Kafka offsets can cause persistent consumer-group lag even when ingestion is low, databases are idle, and no errors are observed.
#spark
fromMedium
2 months ago
Data science

How I Fixed a Critical Spark Production Performance Issue (and Cut Runtime by 70%)

fromMedium
2 months ago
Software development

How I Fixed a Critical Spark Production Performance Issue (and Cut Runtime by 70%)

fromMedium
2 months ago
Data science

How I Fixed a Critical Spark Production Performance Issue (and Cut Runtime by 70%)

fromMedium
2 months ago
Software development

How I Fixed a Critical Spark Production Performance Issue (and Cut Runtime by 70%)

Artificial intelligence
fromInfoQ
2 months ago

[Video Podcast] The Craft of Software Architecture in the Age of AI Tools

Software architecture must be rethought for the age of AI tools, integrating design, platforms, APIs, delivery, and practical experiential guidance for real-world practitioners.
Business intelligence
fromDevOps.com
1 month ago

Why OpenTelemetry Is Paving the Way for the Rise of the Observability Warehouse - DevOps.com

OpenTelemetry adoption drives observability architecture toward unified warehouse models that centralize logs, metrics, and traces for scalable, cost-effective real-time operational intelligence.
Artificial intelligence
fromMedium
2 months ago

Beyond the Monolith: The Rise of the AI Microservices Architecture

LangGraph models AI interactions as a state-machine graph with persistent state, semantic routing, and microservice agents for robust orchestration.
Data science
fromInfoQ
1 month ago

Databricks Introduces Lakebase, a PostgreSQL Database for AI Workloads

Databricks Lakebase is a serverless PostgreSQL OLTP database that separates compute from storage and unifies transactional and analytical capabilities.
Business intelligence
fromTechzine Global
2 months ago

ClickHouse, the open-source challenger to Snowflake and Databricks

ClickHouse is a high-performance columnar OLAP database rapidly adopted by AI and enterprise users, now valued at $15B and acquiring Langfuse.
DevOps
fromInfoQ
1 month ago

Running Ray at Scale on AKS

Microsoft and Anyscale provide guidance for running managed Ray service on Azure Kubernetes Service, addressing GPU capacity limits, ML storage challenges, and credential expiry issues through multi-cluster, multi-region deployment strategies.
Business intelligence
fromNew Relic
2 months ago

Optimize Databricks: Full Visibility with New Relic

New Relic Databricks Integration provides unified telemetry, speeding troubleshooting, improving performance and resource utilization, and linking Databricks performance directly to cost.
DevOps
fromInfoWorld
1 month ago

Running agents with Amazon Bedrock AgentCore

Amazon Bedrock AgentCore provides enterprise-grade infrastructure for deploying and managing AI agents at scale, supporting multiple models, frameworks, and integrations while remaining model-agnostic.
Software development
fromInfoQ
2 months ago

LinkedIn Re-Architects Service Discovery: Replacing Zookeeper with Kafka and xDS at Scale

Moving service discovery from ZooKeeper to a Kafka + xDS-based, eventually consistent architecture enabled scalable, language-agnostic, zero-downtime migration.
fromDevOps.com
1 month ago

Zero Downtime Multicloud Migrations for Observability Control Planes - DevOps.com

An observability control plane isn't just a dashboard. It's the operational authority system. It defines alert rules, routing, ownership, escalation policy, and notification endpoints. When that layer is wrong, the impact is immediate. The wrong team gets paged. The right team never hears about the incident. Your service level indicators look clean while production burns.
DevOps
Software development
fromMedium
2 months ago

The Complete Database Scaling Playbook: From 1 to 10,000 Queries Per Second

Database scaling to 10,000 QPS requires staged architectural strategies timed to traffic thresholds to avoid outages or unnecessary cost.
fromTechzine Global
2 months ago

Databricks makes serverless Postgress service Lakebase available

Databricks today announced the general availability of Lakebase on AWS, a new database architecture that separates compute and storage. The managed serverless Postgres service is designed to help organizations build faster without worrying about infrastructure management. When databases link compute and storage, every query must use the same CPU and memory resources. This can cause a single heavy query to affect all other operations. By separating compute and storage, resources automatically scale with the actual load.
Software development
fromInfoWorld
2 months ago

The 'Super Bowl' standard: Architecting distributed systems for massive concurrency

When I manage infrastructure for major events (whether it is the Olympics, a Premier League match or a season finale) I am dealing with a "thundering herd" problem that few systems ever face. Millions of users log in, browse and hit "play" within the same three-minute window. But this challenge isn't unique to media. It is the same nightmare that keeps e-commerce CTOs awake before Black Friday or financial systems architects up during a market crash. The fundamental problem is always the same: How do you survive when demand exceeds capacity by an order of magnitude?
DevOps
fromDbmaestro
5 years ago

Database Delivery Automation in the Multi-Cloud World

The main advantage of going the Multi-Cloud way is that organizations can "put their eggs in different baskets" and be more versatile in their approach to how they do things. For example, they can mix it up and opt for a cloud-based Platform-as-a-Service (PaaS) solution when it comes to the database, while going the Software-as-a-Service (SaaS) route for their application endeavors.
DevOps
[ Load more ]