#tpc-ds

[ follow ]
Business intelligence
fromDevOps.com
9 hours ago

Dynatrace to Acquire Bindplane to Process and Route Telemetry Data - DevOps.com

Dynatrace is acquiring Bindplane to enhance log management and allow pre-processing of telemetry data for cost reduction and workflow efficiency.
DevOps
fromInfoQ
1 day ago

Google Cloud Highlights Ongoing Work on PostgreSQL Core Capabilities

Google Cloud has made significant technical contributions to PostgreSQL, enhancing logical replication, upgrade processes, and system stability.
Web frameworks
fromInfoQ
1 day ago

Tiger Teams, Evals and Agents: The New AI Engineering Playbook

Sam Bhagwat is a co-founder and CEO of Mastra, an open source JavaScript/Typescript framework for building AI agents.
Data science
fromFast Company
3 days ago

Data, not infrastructure, must drive your AI strategy

Data centricity is essential for effective AI strategies, enabling collaboration and problem-solving across business units by making data accessible.
Artificial intelligence
fromComputerworld
3 days ago

AI often doesn't deliver ROI for IT departments either

Only 28% of AI projects in infrastructure and operations achieve meaningful ROI, with many failing due to unrealistic expectations and skills gaps.
Software development
fromInfoQ
1 week ago

TigerFS Mounts PostgreSQL Databases as a Filesystem for Developers and AI Agents

TigerFS is an experimental filesystem that integrates PostgreSQL, allowing file operations through a standard filesystem interface.
DevOps
fromwww.businessinsider.com
18 hours ago

Amazon creates 'Project Houdini' to make data center delays disappear

Amazon's Project Houdini aims to speed up data center construction by moving processes to factories, addressing AI demand and capacity constraints.
Business intelligence
fromZDNET
1 day ago

I asked 5 data leaders about how they use AI to automate - and end integration nightmares

Strong processes and AI integration are essential for businesses to effectively utilize data.
#snowflake
Django
fromMedium
1 week ago

Snowflake Supports Directory Imports

Easier package imports into Snowflake functions and procedures from stage directories and SnowGit directories streamline development and deployment.
Artificial intelligence
fromTheregister
2 weeks ago

Snowflake's ongoing pitch: bring AI to data, not vice versa

Snowflake is enhancing its platform for AI integration through strategic partnerships and acquisitions, focusing on customer ROI and data management efficiency.
#ai
Data science
fromTheregister
2 weeks ago

Datadog bets DIY AI will mean it dodges the SaaSpocalypse

Datadog is releasing an AI model to enhance its observability tools and mitigate risks from customers building their own solutions.
Data science
fromTheregister
2 weeks ago

Datadog bets DIY AI will mean it dodges the SaaSpocalypse

Datadog is releasing an AI model to enhance its observability tools and mitigate risks from customers building their own solutions.
Marketing tech
fromEMARKETER
1 week ago

Brands want personalization at scale, but their data stack keeps getting in the way

Limited platform integration is the top barrier to personalization for 42% of brand marketers and 47% of agency marketers in North America.
fromSilicon Canals
3 days ago

When militaries share data centers with banks: how Gulf strikes exposed a structural flaw in global cloud infrastructure - Silicon Canals

When civilian banks, logistics platforms, and payment processors share physical data center infrastructure with military AI systems, those facilities become legitimate military targets under international humanitarian law - and the civilian services housed inside lose their legal protection.
Information security
DevOps
fromTheregister
1 day ago

AWS put a file system on S3; I stress-tested it

AWS S3 Files allows mounting S3 buckets as NFS shares, providing solid conflict resolution and cost-effective storage options.
#artificial-intelligence
Artificial intelligence
fromDigiday
4 days ago

AI talk at retail events shifts to proving real results, defining a true strategy

AI has evolved from experimentation to a focus on proven strategies and increased productivity in retail.
Artificial intelligence
fromDigiday
4 days ago

AI talk at retail events shifts to proving real results, defining a true strategy

AI has evolved from experimentation to a focus on proven strategies and increased productivity in retail.
Information security
fromTNW | Insights
6 days ago

KeeperDB brings zero-trust database access to privileged access management

Database credentials are a major attack vector, and KeeperDB integrates access controls into its PAM platform to enhance security.
Science
fromNature
2 weeks ago

Drowning in data sets? Here's how to cut them down to size

The Square Kilometre Array Observatory will generate massive data, but storage and retention pose significant challenges for researchers.
DevOps
fromDevOps.com
3 days ago

Apica Extends Scope and Reach of Platform for Managing Telemetry Data - DevOps.com

Apica's Ascent platform update enhances telemetry data management for DevOps teams, improving observability and cost control.
Java
fromMedium
2 weeks ago

Spark Internals: Understanding Tungsten (Part 1)

Apache Spark revolutionized big data processing but faces challenges due to JVM memory management and garbage collection issues.
DevOps
fromInfoQ
1 week ago

Replacing Database Sequences at Scale Without Breaking 100+ Services

Validating requirements can simplify complex problems, and embedding sequence generation reduces network calls, enhancing performance and reliability.
#databricks
Information security
fromInfoWorld
2 weeks ago

Databricks pitches Lakewatch as a cheaper SIEM - but is it really?

Translating benefits into buy-in from CIOs and CISOs may be challenging for Databricks despite its intent and acquisitions.
Information security
fromInfoWorld
2 weeks ago

Databricks pitches Lakewatch as a cheaper SIEM - but is it really?

Translating benefits into buy-in from CIOs and CISOs may be challenging for Databricks despite its intent and acquisitions.
#observability
DevOps
fromTechzine Global
1 week ago

Observability warehouses, the next structural evolution for telemetry

Observability is essential for real-time insights in cloud systems, helping to reduce downtime and improve performance.
DevOps
fromTechzine Global
1 week ago

Observability warehouses, the next structural evolution for telemetry

Observability is essential for real-time insights in cloud systems, helping to reduce downtime and improve performance.
Data science
fromMedium
3 weeks ago

Building Consistent Data Foundations at Scale

Building consistent data foundations through intentional architecture, engineering, and governance is essential to prevent fragmentation, support AI adoption, ensure regulatory compliance, and enable reliable organizational decisions at scale.
Business intelligence
fromTheregister
1 week ago

Microsoft Fabric Database Hub dubbed 'partial' solution

Microsoft's Fabric Database Hub offers a centralized management solution for its database services but lacks support for non-Microsoft databases.
Software development
fromTechzine Global
4 weeks ago

The RAMpocalypse is a warning for stricter performance KPIs

Rising hardware costs force developers to optimize software efficiency rather than relying on throwing more resources at performance problems.
Information security
fromTechzine Global
2 weeks ago

Databricks launches Lakewatch: agentic SIEM on the Lakehouse

Lakewatch is an open SIEM platform that consolidates security, IT, and business data, enabling rapid threat detection and response using AI agents.
Data science
fromMedium
1 month ago

Migrating to the Lakehouse Without the Big Bang: An Incremental Approach

Query federation enables safe, incremental lakehouse migration by allowing simultaneous queries across legacy warehouses and new lakehouse systems without risky big bang cutover approaches.
DevOps
fromInfoWorld
1 week ago

How to build an enterprise-grade MCP registry

MCP registries are essential for integrating AI agents with enterprise systems, requiring semantic discovery, governance, and developer-friendly controls.
Marketing tech
fromMarTech
1 month ago

The marketing data most companies still fail to measure | MarTech

Phone conversations are critical first-party data sources for marketing measurement as privacy changes weaken traditional attribution signals, making conversation intelligence platforms essential measurement infrastructure.
Software development
fromMedium
1 month ago

Unified Databricks Repository for Scala and Python Data Pipelines

Databricks repositories require structured setup with Gradle for multi-language support, dependency management, and version control to scale beyond manual notebook maintenance.
Marketing
fromSkift Meetings
1 month ago

How to Make Event Data Matter in the Boardroom

Corporate events require data-driven measurement systems connecting to business outcomes to justify budgets and earn strategic credibility with executive leadership.
DevOps
fromInfoQ
2 weeks ago

AWS Expands Aurora DSQL with Playground, New Tool Integrations, and Driver Connectors

Amazon Aurora DSQL introduces usability enhancements, including a browser-based playground and integrations with popular SQL tools for improved developer experience.
Business intelligence
fromInfoWorld
3 weeks ago

Snowflake's new 'autonomous' AI layer aims to do the work, not just answer questions

Project SnowWork is Snowflake's autonomous AI layer that automates data analysis tasks like forecasting, churn analysis, and report generation without requiring data team intervention.
Data science
fromMedium
1 month ago

100 Scala Interview Questions and Answers for Data Engineers

Structured Scala and Apache Spark interview preparation requires understanding distributed systems, performance trade-offs, and pipeline design beyond theoretical knowledge.
#data-center-modernization
DevOps
fromInfoWorld
3 weeks ago

Update your databases now to avoid data debt

Multiple major open source databases reach end-of-life in 2026, requiring teams to plan upgrades and migrations to avoid security risks and higher costs.
Artificial intelligence
fromInfoWorld
1 month ago

Why AI requires rethinking the storage-compute divide

AI workloads require continuous processing of unstructured multimodal data, causing redundant data movement and transformation that wastes infrastructure costs and data scientist time.
Data science
fromInfoWorld
1 month ago

The revenge of SQL: How a 50-year-old language reinvents itself

SQL has experienced a major comeback driven by SQLite in browsers, improved language tools, and PostgreSQL's jsonb type, making it both traditional and exciting for modern development.
Business intelligence
fromEntrepreneur
4 weeks ago

The Game-Changing Tech Saving Companies From Data Disasters

Combining Continuous Data Protection with AI capabilities enables businesses to achieve near-zero Recovery Point Objectives and minimal Recovery Time Objectives, preventing data loss and minimizing downtime.
DevOps
fromTechzine Global
4 weeks ago

Everpure brings ActiveCluster to file environments

Everpure expands its Enterprise Data Cloud platform with ActiveCluster for file environments, enabling seamless data movement between systems while maintaining availability and protecting unstructured data critical for AI applications.
fromTechzine Global
1 month ago

DataBahn and Microsoft accelerate SIEM deployment through integration

DataBahn's AI-driven connectors automatically normalize, enrich, and route telemetry from more than 500 sources to Microsoft Sentinel. DataBahn's Cruz AI engine determines which data to send to the analytics tier and which to the Sentinel data lake for long-term storage. Customers report cost savings of up to 60 percent on Sentinel ingestion thanks to this intelligent tiering mechanism.
Business intelligence
Artificial intelligence
fromEngadget
1 month ago

AI data centers could reduce power draw on demand, study says

AI data centers can dynamically reduce energy consumption by up to 40% without disrupting critical workloads, enabling grid stability and reducing infrastructure strain.
Business intelligence
fromTechzine Global
1 month ago

Dataiku introduces platform for scalable enterprise AI

Dataiku launches Platform for AI Success with three new products designed to move AI initiatives from pilots to measurable business outcomes through unified orchestration across cloud providers.
Miscellaneous
fromComputerWeekly.com
1 month ago

Datacentre demand is huge but power and skills hold things back, survey shows | Computer Weekly

Demand for European datacentres outstrips supply due to power constraints, severe skills shortages, supply‑chain volatility, rising costs, and facilities often unready for AI deployments.
Startup companies
fromInfoQ
2 months ago

Etleap Launches Iceberg Pipeline Platform to Simplify Enterprise Adoption of Apache Iceberg

Managed Iceberg pipeline platform unifies ingestion, transformation, orchestration, and table operations inside customers' VPCs, enabling enterprise Iceberg adoption without building custom stacks.
Data science
fromInfoWorld
1 month ago

Buyer's guide: Comparing the leading cloud data platforms

Five leading cloud data platforms—Databricks, Snowflake, Amazon RedShift, Google BigQuery, and Microsoft Fabric—offer distinct architectural approaches for enterprise data storage, analytics, and AI workloads.
DevOps
fromInfoQ
1 month ago

Google BigQuery Previews Cross-Region SQL Queries for Distributed Data

BigQuery's global queries feature enables SQL queries across multiple geographic regions without data movement, eliminating ETL pipelines for distributed analytics.
fromThe Drum
2 months ago

Deeper data delivers more inspired partnership decisions

Imagine you're selecting an influencer to work with on your new campaign. You've narrowed it down to two, both in the right area, both creating the right sort of content. One has 24.6 million subscribers, the other 1.4 million. Which do you choose? Now imagine you could find out the first had 8.7 million unique viewers last month, while the second had 9.9 million. Do you want to change your mind?
Marketing
fromInfoWorld
2 months ago

AI is changing the way we think about databases

Developers have spent the past decade trying to forget databases exist. Not literally, of course. We still store petabytes. But for the average developer, the database became an implementation detail; an essential but staid utility layer we worked hard not to think about. We abstracted it behind object-relational mappers (ORM). We wrapped it in APIs. We stuffed semi-structured objects into columns and told ourselves it was flexible.
Software development
fromArs Technica
1 month ago

Most VMware users still "actively reducing their VMware footprint," survey finds

CloudBolt's survey also examined how respondents are migrating workloads off of VMware. Currently, 36 percent of participants said they migrated 1-24 percent of their environment off of VMware. Another 32 percent said that they have migrated 25-49 percent; 10 percent said that they've migrated 50-74 percent of workloads; and 2 percent have migrated 75 percent or more of workloads. Five percent of respondents said that they have not migrated from VMware at all.
Tech industry
Data science
fromInfoQ
1 month ago

Pinterest's CDC-Powered Ingestion Slashes Database Latency from 24 Hours to 15 Minutes

Pinterest deployed a next-generation database ingestion framework using CDC, Kafka, Flink, Spark, and Iceberg to reduce data latency from 24+ hours to minutes while processing only changed records.
fromTechzine Global
2 months ago

4 steps to create a future-proof data infrastructure

A future-proof IT infrastructure is often positioned as a universal solution that can withstand any change. However, such a solution does not exist. Nevertheless, future-proofing is an important concept for IT leaders navigating continuous technological developments and security risks, all while ensuring that daily business operations continue. The challenge is finding a balance between reactive problem solving and proactive planning, because overlooking a change can cost your organization. So, how do you successfully prepare for the future without that one-size-fits-all solution?
Tech industry
fromMedium
2 months ago

How I Fixed a Critical Spark Production Performance Issue (and Cut Runtime by 70%)

"The job didn't fail. It just... never finished." That was the worst part. No errors.No stack traces.Just a Spark job running forever in production - blocking downstream pipelines, delaying reports, and waking up-on-call engineers at 2 AM. This is the story of how I diagnosed a real Spark performance issue in production and fixed it drastically, not by adding more machines - but by understanding Spark properly.
Information security
fromSecuritymagazine
2 months ago

Product Spotlight on Analytics

Taelor Sutherland is Associate Editor at Security magazine covering enterprise security, coordinating digital content, and holding a BA in English Literature from Agnes Scott College.
fromTechzine Global
2 months ago

Databricks makes serverless Postgress service Lakebase available

Databricks today announced the general availability of Lakebase on AWS, a new database architecture that separates compute and storage. The managed serverless Postgres service is designed to help organizations build faster without worrying about infrastructure management. When databases link compute and storage, every query must use the same CPU and memory resources. This can cause a single heavy query to affect all other operations. By separating compute and storage, resources automatically scale with the actual load.
Software development
fromFast Company
1 month ago

Beware of data hubris

Organizations are drowning in dashboards, KPIs, performance metrics, behavioral traces, biometric indicators, predictive scores, engagement rates, and AI-generated forecasts. We have more data than we know what to do with. We pretend that the mere presence of data guarantees clarity. It does not. That's data hubris—the arrogant belief that because something can be measured, it can be mastered.
Business intelligence
Data science
fromInfoQ
1 month ago

Databricks Introduces Lakebase, a PostgreSQL Database for AI Workloads

Databricks Lakebase is a serverless PostgreSQL OLTP database that separates compute from storage and unifies transactional and analytical capabilities.
Software development
fromMedium
1 month ago

The Complete Database Scaling Playbook: From 1 to 10,000 Queries Per Second

Database scaling to 10,000 QPS requires staged architectural strategies timed to traffic thresholds to avoid outages or unnecessary cost.
Artificial intelligence
fromBusiness Insider
1 month ago

PwC engineers built an AI agent to tackle the corporate world's least sexy task: spreadsheets

PwC developed a frontier AI agent that reasons across massive, enterprise-grade spreadsheets, addressing complexity, interlinked workbooks, and data dependencies conventional AI struggles with.
Data science
fromInfoQ
2 months ago

Beyond the Warehouse: Why BigQuery Alone Won't Solve Your Data Problems

Data warehouses like BigQuery perform well initially but become slow, costly, and disorganized at scale, undermining low-latency operational use and innovation.
Artificial intelligence
fromInfoWorld
2 months ago

Teradata unveils enterprise AgentStack to push AI agents into production

Teradata positions Enterprise AgentStack as a vendor-agnostic execution layer across hybrid environments, contrasting platform-tied AI approaches from Snowflake and Databricks.
fromNew Relic
3 months ago

The Power and Cost of Data Cardinality

The more attributes you add to your metrics, the more complex and valuable questions you can answer. Every additional attribute provides a new dimension for analysis and troubleshooting. For instance, adding an infrastructure attribute, such as region can help you determine if a performance issue is isolated to a specific geographic area or is widespread. Similarly, adding business context, like a store location attribute for an e-commerce platform, allows you to understand if an issue is specific to a particular set of stores
Data science
Data science
fromMedium
3 months ago

The Complete Guide to Optimizing Apache Spark Jobs: From Basics to Production-Ready Performance

Optimize Spark jobs by using lazy evaluation awareness, early filter and column pruning, partition pruning, and appropriate join strategies to minimize shuffles and I/O.
fromTechzine Global
1 month ago

Databricks shows how AI strengthens the SaaS model

The rise of generative AI is often seen as an existential threat to the SaaS model. Interfaces would disappear, software would fade away, and existing players would become irrelevant. However, new figures from Databricks paint a different picture. Rather than undermining SaaS, AI appears to be increasing its use. This week, Databricks reported a revenue run rate of $5.4 billion, a 65 percent year-on-year increase. More than a quarter of that now comes from AI-related products.
Artificial intelligence
Data science
fromComputerworld
2 months ago

Tableau re-engineers dashboards, adds new analytics tools for business analysts

Tableau 2022.3 adds Data Guide and Table Extension, dynamic dashboards, event auditing, and performance/cost optimization to simplify self-service analytics for business users.
Business intelligence
fromTechzine Global
2 months ago

ClickHouse, the open-source challenger to Snowflake and Databricks

ClickHouse is a high-performance columnar OLAP database rapidly adopted by AI and enterprise users, now valued at $15B and acquiring Langfuse.
Artificial intelligence
fromTechzine Global
2 months ago

Snowflake launches Cortex Code agent for understanding data context

Cortex Code is an AI agent that converts complex data engineering, ML, and analytics tasks into natural-language workflows integrated into Snowflake and developer tools.
Data science
fromCIO
2 months ago

5 perspectives on modern data analytics

Data/business analytics is the top IT investment priority, yet analytics projects often fail due to poor data, vague objectives, and one-size-fits-all solutions.
fromDbmaestro
5 years ago

Database Delivery Automation in the Multi-Cloud World

The main advantage of going the Multi-Cloud way is that organizations can "put their eggs in different baskets" and be more versatile in their approach to how they do things. For example, they can mix it up and opt for a cloud-based Platform-as-a-Service (PaaS) solution when it comes to the database, while going the Software-as-a-Service (SaaS) route for their application endeavors.
DevOps
fromUX Magazine
1 month ago

What "Cheap Prediction" Means for Enterprise

Many organizations, Gans suggests, resemble public airports - full of people waiting for phones to ring, managing buffers, absorbing uncertainty.
Artificial intelligence
Data science
fromInfoWorld
2 months ago

Snowflake debuts Cortex Code, an AI agent that understands enterprise data context

Cortex Code enables developers to use natural language to build, optimize, and deploy governed, production-ready data pipelines, analytics, ML workloads, and AI agents.
fromTechRepublic
2 months ago

What Are the Pros and Cons of Data Centers?

When ChatGPT launched in late 2022, I watched something remarkable happen. Within two months, it hit 100 million users, a growth rate that sent shockwaves through Silicon Valley. Today, it has over 800 million weekly active users. That launch sparked an explosion in AI development that has fundamentally changed how we build and operate the infrastructure powering our digital world.
Artificial intelligence
Artificial intelligence
fromTechzine Global
2 months ago

IBM launches Enterprise Advantage for scaling AI

IBM Enterprise Advantage helps organizations rapidly build and scale custom AI platforms across multiple clouds and models using IBM consulting expertise and existing infrastructure.
Artificial intelligence
fromInfoWorld
1 month ago

Five MCP servers to rule the cloud

Major cloud providers now offer official MCP servers that let AI agents automate cloud operations using existing cloud credentials and natural language commands.
[ Load more ]