Futurology articles
Ventiva shows off tech to keep chips and devices cool
In context: We've noticed a burning sensation in our pockets lately. Not the bite of inflation, but an actual heat source coming from our phones. Maybe it is 5G, maybe it is a design decision made by Apple, or by TSMC, whatever the reason, our smartphone runs hot.
Eco-friendly "water batteries" are cheaper and safer than lithium-ions
Goodbye to Graphics: How GPUs Came to Dominate Compute and AI
Gone are the days when the sole function for a graphics chip were, graphics. Let's explore how the GPU evolved from a modest pixel pusher into a blazing powerhouse of floating-point computation.
Data Governance and Provenance: Two words that are critical to the future of generative AI
DVD-like optical disc could store 1.6 petabits (or 200 terabytes) on 100 layers
Wrestling with AI and the AIpocalypse we should be worried about
Editor's take: Like almost everyone in tech today, we have spent the past year trying to wrap our heads around "AI". What it is, how it works, and what it means for the industry. We are not sure that we have any good answers, but a few things have been clear. Maybe AGI (artificial general intelligence) will emerge, or we'll see some other major AI breakthrough, but focusing too much on those risks could be overlooking the very real – but also very mundane – improvements that transformer networks are already delivering.
Custom data center reaches 64 gigatransfers per second through optical PCIe 6.0
HaLow Wi-Fi standard achieves 1.8-mile range in field test
Researchers develop world's first functioning graphene semiconductor
LG shows off 77-inch transparent OLED TV at CES
A new internet standard called L4S aims to lower latency, bring more speed to the web
Researchers achieve data storage breakthrough using diamond defects
Who famously said: "There is no reason anyone would want a computer in their home"?
Opinion: The rapidly evolving state of Generative AI
SoftBank CEO says artificial general intelligence (AGI) will be here in 10 years
Testing Nvidia DLSS 3.5 Ray Reconstruction Using Cyberpunk 2.0
The next iteration of Nvidia's DLSS technology has landed and today we're checking out what DLSS 3.5 is and taking a closer look at ray reconstruction, which has been integrated into Cyberpunk 2.0.
New silicon-based "metasurfaces" can rapidly detect toxins, viruses, cancer, and other diseases
Then and Now: How 30 Years of Progress Have Changed PCs
Over the last 30 years, the PC has utterly transformed in appearance, capability, and usage. From the hulking beige boxes to an astonishing array of powerful, colorful, and astonishing computers.
I got to play with the Apple Vision Pro and saw the future of computing. Again.
Editor's take: I was one of the lucky few who got to attend Apple's WWDC keynote presentation in person, and also got to try the new Apple Vision Pro headset for a 25-minute hands-on, er, heads-on demo. The experience was very good – as it certainly should be for a product that's going to cost a whopping $3,499 – but it was also a bit more similar to other devices I've tried over the years than I initially expected it to be.
What Are Chiplets and Why They Are So Important for the Future of Processors
While chiplets have been in use for decades, today they are the hottest trend in processor tech and at the cutting edge of technology, with millions of people worldwide using them in desktop PCs, workstations, and servers.
Experimental AI tool lets you morph images with a simple click and drag workflow
Nvidia's ambitious future is to bring AI to every industry where GPU tech can be leveraged
The State of Self-Driving Cars: Autonomous Advances
Significant advances have been made in the technology that powers self-driving cars in the last decade, but there's still further to go. We dive into the current state of self-driving cars in 2023.