Neo4j Aura Agent is an end-to-end platform for creating agents, connecting them to knowledge graphs, and deploying to production in minutes. In this post, we'll explore the features of Neo4j Aura Agent that make this all possible, along with links to coded examples to get hands-on with the platform.
Our customers, ranging from large enterprises to AI research labs, are no longer just asking for AI features. They need a way to collect high-fidelity, synchronized robot and vision data to train AI models on the same robots they intend to deploy. Our AI Trainer is the industry's first direct lab-to-factory solution for AI model training.
Since its release in 2021, this repository has become a bedrock in discovery and a first port of call for research projects that try to understand life at the molecular level. But previous iterations of the database lacked predictions of how proteins form complexes, which can be indispensable for their function.
Google tasked Gemini with sorting through 5 million news articles from around the world and isolating flood reports. It transformed this data into a geo-tagged series of chronological events. Next, researchers trained a model to ingest current weather forecasts and leverage the Groundsource data to determine the likelihood of a flash flood in a given area.
A comprehensive analysis of Google search patterns related to birds explores what species people seek information about most frequently. The investigation spans six interconnected analyses examining bird variety, taxonomic classifications, information sharing behaviors, birder sighting correlations with search trends, regional popularity differences across states, and temporal patterns in search interest.
Hedge funds and other money managers spent $2.8 billion on alternative data in 2025, according to a new report from consultancy Neudata, a 17% jump from the year before. It's more than double what asset managers spent on alternative data in 2021, which includes a wide range of non-traditional information sources. The report projects that the total spend on alternative datasets could jump to more than $23 billion in the consultancy's bull case in 2030 and just under $8 billion in the bear case.
AI was everywhere, but I wasn't focused on product launches. I was looking at how companies think about data itself: how it's shared, governed and ultimately turned into decisions. And across conversations with executives and sessions on security and compliance, a pattern emerged: the technical limitations that once justified locking data down have largely been solved. What remains difficult is human. Alignment, trust and confidence inside organizations are now the true barriers.
Every search, purchase, loyalty swipe, location ping and scroll feeds systems that now shape pricing, product decisions, hiring and marketing strategies. Most founders understand this in theory, but few grasp the practical consequence: whether they intend to or not, they and their customers are already casting votes with their data. And those votes? They're usually cast passively, on someone else's terms.
I wrote a book for O'Reilly on scaling machine learning with Spark specifically. My second book is coming out on how to improve high-performance Spark, the second edition. Started my career in the machine learning space 15 years ago, moved into data infrastructure, batch processing, and a year and a half ago I moved into the data streaming space, which I think it's what's going to help us pave the future in the data.
Most beginner data portfolios look similar. They include: A few cleaned datasets Some charts or dashboards A notebook with code and commentary Again, nothing here is wrong. But hiring teams don't review portfolios to check whether you can follow instructions. They review them to see whether you can think like a data analyst. When projects feel generic, reviewers are left guessing:
Every year, poor communication and siloed data bleed companies of productivity and profit. Research shows U.S. businesses lose up to $1.2 trillion annually to ineffective communication, that's about $12,506 per employee per year. This stems from breakdowns that waste an average of 7.47 hours per employee each week on miscommunications. The damage isn't only interpersonal; it's structural. Disconnected and fragmented data systems mean that employees spend around 12 hours per week just searching for information trapped in those silos.
For a brief moment in October, Alejandro Quintero thought he had made it big in China. The Bogotá-based data analyst owns and manages a website that publishes articles about paranormal activities, like ghosts and aliens. The content is written in "Spanglish," he says, and was never intended for an Asian audience. But last fall, Quintero's site suddenly began receiving a large volume of visits from China and Singapore.