#data-warehouse-transformation

[ follow ]
#snowflake
Django
fromMedium
2 days ago

Snowflake Supports Directory Imports

Easier package imports into Snowflake functions and procedures from stage directories and SnowGit directories streamline development and deployment.
Artificial intelligence
fromTheregister
1 week ago

Snowflake's ongoing pitch: bring AI to data, not vice versa

Snowflake is enhancing its platform for AI integration through strategic partnerships and acquisitions, focusing on customer ROI and data management efficiency.
Django
fromMedium
2 days ago

Snowflake Supports Directory Imports

Easier package imports into Snowflake functions and procedures from stage directories and SnowGit directories streamline development and deployment.
Artificial intelligence
fromTheregister
1 week ago

Snowflake's ongoing pitch: bring AI to data, not vice versa

Snowflake is enhancing its platform for AI integration through strategic partnerships and acquisitions, focusing on customer ROI and data management efficiency.
DevOps
fromInfoQ
1 day ago

Replacing Database Sequences at Scale Without Breaking 100+ Services

Validating requirements can simplify complex problems, and embedding sequence generation reduces network calls, enhancing performance and reliability.
Software development
fromInfoQ
20 hours ago

TigerFS Mounts PostgreSQL Databases as a Filesystem for Developers and AI Agents

TigerFS is an experimental filesystem that integrates PostgreSQL, allowing file operations through a standard filesystem interface.
#enterprise-ai
Business intelligence
fromTechzine Global
1 day ago

All shook up, IFS unlocks asset-based pricing for enterprise AI

IFS introduces an outcomes-based pricing model for enterprise AI, aligning software costs with operational assets instead of user counts.
Business intelligence
fromTechzine Global
1 day ago

All shook up, IFS unlocks asset-based pricing for enterprise AI

IFS introduces an outcomes-based pricing model for enterprise AI, aligning software costs with operational assets instead of user counts.
Artificial intelligence
fromInfoWorld
1 day ago

Google gives enterprises new controls to manage AI inference costs and reliability

Gemini API introduces Flex and Priority tiers for managing AI inference workloads based on criticality and cost.
Tech industry
fromWIRED
2 days ago

A New Google-Funded Data Center Will Be Powered by a Massive Gas Plant

A pragmatic 'all-of-the-above' strategy is essential for energy, with gas as a critical bridge while investing in renewables.
Marketing tech
fromEMARKETER
4 days ago

Brands want personalization at scale, but their data stack keeps getting in the way

Limited platform integration is the top barrier to personalization for 42% of brand marketers and 47% of agency marketers in North America.
fromInfoWorld
4 days ago

How Apache Kafka flexed to support queues

Apache Kafka has cemented itself as the de facto platform for event streaming, often referred to as the 'universal data substrate' due to its extensive ecosystem that enables connectivity and processing capabilities.
Scala
Business
fromFortune
5 days ago

How Dell reinvented itself as an AI-server powerhouse - and what its CFO is building next | Fortune

Dell Technologies has shifted focus to AI infrastructure, recording significant growth in AI server orders and rethinking finance operations with AI agents.
#databricks
Information security
fromInfoWorld
1 week ago

Databricks pitches Lakewatch as a cheaper SIEM - but is it really?

Translating benefits into buy-in from CIOs and CISOs may be challenging for Databricks despite its intent and acquisitions.
Information security
fromInfoWorld
1 week ago

Databricks pitches Lakewatch as a cheaper SIEM - but is it really?

Translating benefits into buy-in from CIOs and CISOs may be challenging for Databricks despite its intent and acquisitions.
#observability
DevOps
fromTechzine Global
2 days ago

Observability warehouses, the next structural evolution for telemetry

Observability is essential for real-time insights in cloud systems, helping to reduce downtime and improve performance.
DevOps
fromTechzine Global
2 days ago

Observability warehouses, the next structural evolution for telemetry

Observability is essential for real-time insights in cloud systems, helping to reduce downtime and improve performance.
Business intelligence
fromInfoWorld
2 days ago

Kilo targets shadow AI agents with a managed enterprise platform

KiloClaw for Organizations enhances AI agent management with centralized governance, addressing security and compliance concerns for enterprises.
#ai
Data science
fromTheregister
1 week ago

Datadog bets DIY AI will mean it dodges the SaaSpocalypse

Datadog is releasing an AI model to enhance its observability tools and mitigate risks from customers building their own solutions.
Data science
fromTheregister
1 week ago

Datadog bets DIY AI will mean it dodges the SaaSpocalypse

Datadog is releasing an AI model to enhance its observability tools and mitigate risks from customers building their own solutions.
Software development
fromTechzine Global
4 days ago

The ERP that doesn't care which AI you use, and why that's smart

NetSuite announced three new AI Connector Service extensions, emphasizing a strategic shift towards openness and integration with external AI models.
Scala
fromMedium
5 days ago

Data Extraction and Classification Using Structural Pattern Matching in Scala

Scala pattern matching enhances code readability and extensibility in real-world data engineering use cases.
Information security
fromTechzine Global
1 week ago

Databricks launches Lakewatch: agentic SIEM on the Lakehouse

Lakewatch is an open SIEM platform that consolidates security, IT, and business data, enabling rapid threat detection and response using AI agents.
DevOps
fromInfoWorld
4 days ago

Azure's new AI modernization tools

Microsoft's Azure Copilot aids in application migration and modernization, addressing technical debt and improving cloud infrastructure management.
Data science
fromInfoQ
1 week ago

Data Mesh in Action: A Journey From Ideation to Implementation

Data mesh is essential for organizations to develop independent data analytics capabilities after separation from larger parent companies.
Business intelligence
fromTheregister
5 days ago

Microsoft Fabric Database Hub dubbed 'partial' solution

Microsoft's Fabric Database Hub offers a centralized management solution for its database services but lacks support for non-Microsoft databases.
Information security
fromTechzine Global
1 week ago

Google Cloud expands its AI security capabilities through Wiz integration

Google is enhancing security with AI, integrating Wiz for multi-cloud protection and introducing new capabilities to combat evolving cyber threats.
DevOps
fromInfoQ
6 days ago

ProxySQL Introduces Multi-Tier Release Strategy With Stable, Innovative, and AI Tracks

ProxySQL 3.0.6 introduces a multi-tier release strategy focusing on stability, innovation, and AI capabilities for diverse user needs.
Data science
fromMedium
2 weeks ago

Building Consistent Data Foundations at Scale

Building consistent data foundations through intentional architecture, engineering, and governance is essential to prevent fragmentation, support AI adoption, ensure regulatory compliance, and enable reliable organizational decisions at scale.
Business intelligence
fromTheregister
5 days ago

SAP eyes broader data integration with Reltio acquisition

SAP is acquiring Reltio to enhance data integration and support its AI platform with improved master data management capabilities.
Data science
fromMedium
4 weeks ago

Migrating to the Lakehouse Without the Big Bang: An Incremental Approach

Query federation enables safe, incremental lakehouse migration by allowing simultaneous queries across legacy warehouses and new lakehouse systems without risky big bang cutover approaches.
DevOps
fromInfoQ
1 week ago

AWS Expands Aurora DSQL with Playground, New Tool Integrations, and Driver Connectors

Amazon Aurora DSQL introduces usability enhancements, including a browser-based playground and integrations with popular SQL tools for improved developer experience.
#ai-automation
Artificial intelligence
fromTechzine Global
2 weeks ago

Snowflake's Project SnowWork targets autonomous enterprise AI

Snowflake launches Project SnowWork, an autonomous AI interface that performs enterprise tasks like forecasts and reports without data team involvement, expanding from backend infrastructure to front-office productivity tool.
fromTheregister
3 weeks ago
Artificial intelligence

Perplexity: Everything is Computer

Perplexity launches Computer for Enterprise, an AI orchestration service that automates business tasks across integrated cloud applications like Gmail, Slack, and Salesforce.
Artificial intelligence
fromTechzine Global
2 weeks ago

Snowflake's Project SnowWork targets autonomous enterprise AI

Snowflake launches Project SnowWork, an autonomous AI interface that performs enterprise tasks like forecasts and reports without data team involvement, expanding from backend infrastructure to front-office productivity tool.
Artificial intelligence
fromTheregister
3 weeks ago

Perplexity: Everything is Computer

Perplexity launches Computer for Enterprise, an AI orchestration service that automates business tasks across integrated cloud applications like Gmail, Slack, and Salesforce.
Software development
fromMedium
1 month ago

Unified Databricks Repository for Scala and Python Data Pipelines

Databricks repositories require structured setup with Gradle for multi-language support, dependency management, and version control to scale beyond manual notebook maintenance.
DevOps
fromInfoWorld
2 weeks ago

Update your databases now to avoid data debt

Multiple major open source databases reach end-of-life in 2026, requiring teams to plan upgrades and migrations to avoid security risks and higher costs.
Business intelligence
fromInfoWorld
2 weeks ago

Snowflake's new 'autonomous' AI layer aims to do the work, not just answer questions

Project SnowWork is Snowflake's autonomous AI layer that automates data analysis tasks like forecasting, churn analysis, and report generation without requiring data team intervention.
Data science
fromMedium
1 month ago

100 Scala Interview Questions and Answers for Data Engineers

Structured Scala and Apache Spark interview preparation requires understanding distributed systems, performance trade-offs, and pipeline design beyond theoretical knowledge.
DevOps
fromTheregister
2 weeks ago

AWS S3 turns 20 and reaches 'hundreds of exabytes'

Amazon S3 celebrates 20 years of operation, growing from 1 petabyte capacity to storing over 500 trillion objects while maintaining complete API backward compatibility since 2006.
Data science
fromInfoWorld
1 month ago

The revenge of SQL: How a 50-year-old language reinvents itself

SQL has experienced a major comeback driven by SQLite in browsers, improved language tools, and PostgreSQL's jsonb type, making it both traditional and exciting for modern development.
Business intelligence
fromInfoWorld
3 weeks ago

Why Postgres has won as the de facto database: Today and for the agentic future

Leading enterprises achieve 5x ROI by adopting open source databases like PostgreSQL to unify structured and unstructured data for agentic AI, with 81% of successful enterprises committed to open source strategies.
Artificial intelligence
fromInfoWorld
1 month ago

Why AI requires rethinking the storage-compute divide

AI workloads require continuous processing of unstructured multimodal data, causing redundant data movement and transformation that wastes infrastructure costs and data scientist time.
#bigquery
DevOps
fromInfoQ
3 weeks ago

Google BigQuery Previews Cross-Region SQL Queries for Distributed Data

BigQuery's global queries feature enables SQL queries across multiple geographic regions without data movement, eliminating ETL pipelines for distributed analytics.
fromInfoQ
2 months ago
Artificial intelligence

Google BigQuery Adds SQL-Native Managed Inference for Hugging Face Models

DevOps
fromInfoQ
3 weeks ago

Google BigQuery Previews Cross-Region SQL Queries for Distributed Data

BigQuery's global queries feature enables SQL queries across multiple geographic regions without data movement, eliminating ETL pipelines for distributed analytics.
fromInfoQ
2 months ago
Artificial intelligence

Google BigQuery Adds SQL-Native Managed Inference for Hugging Face Models

Business intelligence
fromEntrepreneur
3 weeks ago

The Game-Changing Tech Saving Companies From Data Disasters

Combining Continuous Data Protection with AI capabilities enables businesses to achieve near-zero Recovery Point Objectives and minimal Recovery Time Objectives, preventing data loss and minimizing downtime.
Data science
fromInfoWorld
1 month ago

Buyer's guide: Comparing the leading cloud data platforms

Five leading cloud data platforms—Databricks, Snowflake, Amazon RedShift, Google BigQuery, and Microsoft Fabric—offer distinct architectural approaches for enterprise data storage, analytics, and AI workloads.
Business intelligence
fromTechzine Global
3 weeks ago

Dataiku introduces platform for scalable enterprise AI

Dataiku launches Platform for AI Success with three new products designed to move AI initiatives from pilots to measurable business outcomes through unified orchestration across cloud providers.
Python
fromPyImageSearch
2 months ago

Integrating Streamlit with Snowflake for Live Cloud Data Apps (Part 1) - PyImageSearch

Connect Streamlit to Snowflake to securely run real-time SQL queries, visualize and blend warehouse data with local sources, creating live interactive Python data applications.
Startup companies
fromInfoQ
2 months ago

Etleap Launches Iceberg Pipeline Platform to Simplify Enterprise Adoption of Apache Iceberg

Managed Iceberg pipeline platform unifies ingestion, transformation, orchestration, and table operations inside customers' VPCs, enabling enterprise Iceberg adoption without building custom stacks.
Web development
fromInfoQ
2 months ago

DuckDB's WebAssembly Client Allows Querying Iceberg Datasets in the Browser

DuckDB-Wasm enables browser-based, serverless end-to-end query, read, and write access to Iceberg REST catalogs and object storage without infrastructure setup.
fromDbmaestro
4 years ago

5 Pillars of Database Compliance Automation |

There is a growing emphasis on database compliance today due to the stricter enforcement of compliance rules and regulations to safeguard user privacy. For example, GDPR fines can reach £17.5 million or 4% of annual global turnover (the higher of the two applies). Besides the direct monetary implications, companies also need to prioritize compliance to protect their brand reputation and achieve growth.
EU data protection
Environment
fromTechzine Global
2 months ago

Snowflake launches Energy Solutions for the energy sector

Snowflake Energy Solutions unifies IT, OT, and IoT data, adding governance, partner solutions, and industry datasets to enable AI-driven operations and resilience for energy companies.
fromInfoQ
2 months ago

350PB, Millions of Events, One System: Inside Uber's Cross-Region Data Lake and Disaster Recovery

Uber has built HiveSync, a sharded batch replication system that keeps Hive and HDFS data synchronized across multiple regions, handling millions of Hive events daily. HiveSync ensures cross-region data consistency, enables Uber's disaster recovery strategy, and eliminates inefficiency caused by the secondary region sitting idle, which previously incurred hardware costs equal to the primary, while still maintaining high availability. Built initially on the open-source Airbnb ReAir project, HiveSync has been extended with sharding, DAG-based orchestration, and a separation of control and data planes.
Tech industry
fromMoz
1 month ago

Why Export GA4 Data to BigQuery?

Then coming on to the next point, which is you can create your own sessions and user properties. Now you can do this in the GA4 interface under Explorations.
Marketing tech
Data science
fromInfoQ
1 month ago

Databricks Introduces Lakebase, a PostgreSQL Database for AI Workloads

Databricks Lakebase is a serverless PostgreSQL OLTP database that separates compute from storage and unifies transactional and analytical capabilities.
fromInfoWorld
2 months ago

AI is changing the way we think about databases

Developers have spent the past decade trying to forget databases exist. Not literally, of course. We still store petabytes. But for the average developer, the database became an implementation detail; an essential but staid utility layer we worked hard not to think about. We abstracted it behind object-relational mappers (ORM). We wrapped it in APIs. We stuffed semi-structured objects into columns and told ourselves it was flexible.
Software development
Tech industry
fromTheregister
1 month ago

Snowflake plugs PostgreSQL into its AI Data Cloud

Snowflake now offers a native PostgreSQL DBaaS in its AI Data Cloud to run transactional workloads alongside analytics and AI under unified governance.
fromMarTech
2 months ago

The data quality paradigm shift has arrived | MarTech

If you work in martech, marketing operations or related roles, you've surely heard colleagues and leadership complaining about data quality and their lack of trust in data. We often place the blame for data quality on the system, because we're not willing to fully say the quiet part out loud: The No. 1 factor in data quality is the people, the processes and the level of rigor in those processes.
Marketing tech
fromTechzine Global
2 months ago

Sumo Logic launches data pipeline apps for Snowflake and Databricks

Snowflake offers a fully managed data platform, but Sumo Logic users often lack insight into performance, login activity, and operational health. The Sumo Logic Snowflake Logs App analyzes login and access activity to identify anomalies or suspicious behavior. It also optimizes data pipelines with insights into long-running or failing queries. Teams can centralize log data to facilitate correlation across applications, cloud services, and data platforms.
Information security
fromTechzine Global
2 months ago

4 steps to create a future-proof data infrastructure

A future-proof IT infrastructure is often positioned as a universal solution that can withstand any change. However, such a solution does not exist. Nevertheless, future-proofing is an important concept for IT leaders navigating continuous technological developments and security risks, all while ensuring that daily business operations continue. The challenge is finding a balance between reactive problem solving and proactive planning, because overlooking a change can cost your organization. So, how do you successfully prepare for the future without that one-size-fits-all solution?
Tech industry
fromTechzine Global
1 month ago

Databricks makes serverless Postgress service Lakebase available

Databricks today announced the general availability of Lakebase on AWS, a new database architecture that separates compute and storage. The managed serverless Postgres service is designed to help organizations build faster without worrying about infrastructure management. When databases link compute and storage, every query must use the same CPU and memory resources. This can cause a single heavy query to affect all other operations. By separating compute and storage, resources automatically scale with the actual load.
Software development
Information security
fromSecuritymagazine
1 month ago

Product Spotlight on Analytics

Taelor Sutherland is Associate Editor at Security magazine covering enterprise security, coordinating digital content, and holding a BA in English Literature from Agnes Scott College.
Software development
fromInfoQ
2 months ago

AWS Adds Intelligent-Tiering and Replication for S3 Tables

S3 Tables now support Intelligent-Tiering automatic cost optimization and cross-region/account Apache Iceberg table replication without manual synchronization.
Data science
fromInfoQ
1 month ago

Beyond the Warehouse: Why BigQuery Alone Won't Solve Your Data Problems

Data warehouses like BigQuery perform well initially but become slow, costly, and disorganized at scale, undermining low-latency operational use and innovation.
#dynamodb
#generative-ai
Data science
fromInfoWorld
1 month ago

Snowflake debuts Cortex Code, an AI agent that understands enterprise data context

Cortex Code enables developers to use natural language to build, optimize, and deploy governed, production-ready data pipelines, analytics, ML workloads, and AI agents.
Software development
fromInfoWorld
2 months ago

Why your next microservices should be streaming SQL-driven

Streaming SQL with UDFs, materialized results, and ML/AI integrations enables continuous, stateful processing of event streams for microservices.
fromDbmaestro
5 years ago

Database Delivery Automation in the Multi-Cloud World

The main advantage of going the Multi-Cloud way is that organizations can "put their eggs in different baskets" and be more versatile in their approach to how they do things. For example, they can mix it up and opt for a cloud-based Platform-as-a-Service (PaaS) solution when it comes to the database, while going the Software-as-a-Service (SaaS) route for their application endeavors.
DevOps
Data science
fromMedium
2 months ago

Migrating from Historical Batch Processing to Incremental CDC Using Apache Iceberg (Glue 4...

Use Apache Iceberg Copy-on-Write tables in AWS Glue 4 to migrate from full historical batch reprocessing to incremental CDC, reducing redundant computation, I/O, and costs.
Business intelligence
fromTechzine Global
2 months ago

ClickHouse, the open-source challenger to Snowflake and Databricks

ClickHouse is a high-performance columnar OLAP database rapidly adopted by AI and enterprise users, now valued at $15B and acquiring Langfuse.
Artificial intelligence
fromTechzine Global
1 month ago

Snowflake launches Cortex Code agent for understanding data context

Cortex Code is an AI agent that converts complex data engineering, ML, and analytics tasks into natural-language workflows integrated into Snowflake and developer tools.
Data science
fromCIO
2 months ago

5 perspectives on modern data analytics

Data/business analytics is the top IT investment priority, yet analytics projects often fail due to poor data, vague objectives, and one-size-fits-all solutions.
Data science
fromMedium
2 months ago

The Complete Guide to Optimizing Apache Spark Jobs: From Basics to Production-Ready Performance

Optimize Spark jobs by using lazy evaluation awareness, early filter and column pruning, partition pruning, and appropriate join strategies to minimize shuffles and I/O.
Artificial intelligence
fromTechzine Global
1 month ago

Snowflake CEO: Software risks becoming a "dumb data pipe" for AI

Centralized AI access to enterprise data risks reducing business applications to mere data pipes unless applications deliver clear added value in accuracy, security, and usability.
Artificial intelligence
fromTechRepublic
6 months ago

New AI Data 'Universal Translator' From Salesforce, Snowflake, Others

Snowflake and other firms created the Open Semantic Interchange to standardize semantics and enable interoperable data sharing among AI-enabled products, reducing semantic mismatches.
Data science
fromDevOps.com
1 month ago

Why Data Contracts Need Apache Kafka and Apache Flink - DevOps.com

Data contracts formalize schemas, types, and quality constraints through early producer-consumer collaboration to prevent pipeline failures and reduce operational downtime.
Data science
fromComputerworld
2 months ago

Tableau re-engineers dashboards, adds new analytics tools for business analysts

Tableau 2022.3 adds Data Guide and Table Extension, dynamic dashboards, event auditing, and performance/cost optimization to simplify self-service analytics for business users.
Artificial intelligence
fromMedium
2 months ago

Extracting AI-Ready Data From Organizational Documents

Poor document extraction corrupts retrieval; preserving document structure at ingestion produces reliable embeddings and trustworthy RAG outputs.
Artificial intelligence
fromInfoWorld
2 months ago

Generative AI and the future of databases

Databases must evolve into AI-native systems that securely federate with LLMs, support real-time access, granular permissions, and tools for safe natural-language-to-SQL integration.
[ Load more ]