Home

Technology is the sum of techniques, skills, methods, and processes used in the production of goods or services or in the accomplishment of objectives, such as scientific investigation. Technology can be the knowledge of techniques, processes, and the like, or it can be embedded in machines to allow for operation without detailed knowledge of their workings.
Technologies: Cell phones, computers, video games, televisions, headphones, printers, wearables, musical instruments, home audio, and software. #ad

Gizmodo



Lifehacker



Google



CNET



Android Authority



AppleInsider

  • Red Bull's 2026 Formula 1 launch puts Apple hardware front and center Fri, 16 Jan 2026 18:28:04 +0000
    Red Bull's 2026 Formula 1 launch made Apple's expanding role around the sport hard to ignore, from iPhone-shot footage embedded in the broadcast to Apple Vision Pro wearers taking the stage.

    People on stage present wearing Apple Vision Pro, a label in the corner of the image says 'shot on iPhone'
    Red Bull's F1 launch was full of Apple promotion

    Oracle Red Bull Racing and Visa Cash App Racing Bulls launched their 2026 cars together. The event was streamed live on Red Bull's platforms, marking the start of the Red Bull Ford Powertrains era.

    It wasn't a race broadcast, but a polished Formula 1 media moment for a global audience.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Apple's workhorse MacBook Air is 18 years old Fri, 16 Jan 2026 18:01:08 +0000
    The current Apple Silicon MacBook Air is barely recognizable from the original in all but name, but it's still the same superb line that was unveiled this week in 2008 in a memorable Steve Jobs presentation.

    Thin silver laptop with a glowing logo, shown from the side, casting a dark shadow on a white background.
    Apple advertised the original MacBook Air as "thinpossible" — image credit: Apple

    Cast your mind back to the last year of the George W. Bush Administration. Back then, the phone you really wanted was the iPhone 3G, and the MacBook you were considering was a whole 1.08 inches thick.

    At that time — despite all the efforts of Microsoft Office and Lotus Notes — it was also still common to get internal mail in corporations, too. If you were in a big enough company at the time, you were familiar with those inter-office mail envelopes.


    Continue Reading on AppleInsider | Discuss on our Forums
  • In case you ever doubted it, Apple Car was real -- reveals Airbnb Fri, 16 Jan 2026 17:12:32 +0000
    You could have studied literally thousands of Apple patents, you could have counted the permits it's had to be granted, or could just have waited for Airbnb to show us that Apple really, truly, was working on an Apple Car.

    Digital car dashboard showing 60 mph speed, 200 miles fuel range, lane-keeping graphic with nearby vehicles, 3rd gear in Sport mode, 72F temperature, and 195F engine temperature within a steering wheel view
    Apple Car Ultra is here, but you know this is how an actual Apple Car would have worked — image credit: Apple

    Maybe today "Apple Car" keeps getting autocompleted to "Apple Card," but you know the car was a real thing. Apple still refuses to say a word, yet there's never been a doubt that Apple Car existed. Not since about 2012 when AppleInsider's Daniel Eran Dilger saw the signs and detailed the reasons why Apple had to be doing this.

    Flash forward to now, when the project has been abandoned and still Apple will not say a word about the Apple Car. But it doesn't need to, since Airbnb has spilled the beans in a new hire's biography.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Apple's 1TB M5 MacBook Pro hits $1,599 with new B&H Photo deal Fri, 16 Jan 2026 16:56:15 +0000
    B&H Photo has kicked off the weekend with a standout MacBook Pro deal, dropping the M5 14-inch model with a 1TB SSD to $1,599, the lowest price available for this configuration.

    Silver MacBook Pro 14-inch with Apple logo on a desk, flanked by Apple TV boxes, orange smart speaker, glowing blue spherical lamp, and framed AppleInsider poster against a brick wall
    Save $200 on Apple's 1TB M5 MacBook Pro at B&H.

    January MacBook Pro discounts are in full swing, with B&H Photo dropping the M5 14-inch MacBook Pro with 16GB of RAM and a 1TB SSD to $1,599 when ordered in silver, the most aggressive price we've tracked for this config in 2026.

    Buy 1TB M5 MacBook Pro for $1,599


    Continue Reading on AppleInsider | Discuss on our Forums
  • Apple hit with $150K fine after breaking decade-old agreement in NJ Fri, 16 Jan 2026 16:28:45 +0000
    Apple must change how it displays prices in New Jersey stores after regulators said the company repeatedly failed to comply with pricing requirements it agreed to follow in 2017.

    Spacious modern Apple store interior with large glowing Apple logo, sleek gray walls, bright ceiling lights, and rows of wooden tables displaying neatly arranged laptops, tablets, and phonesNew Jersey fined Apple over repeat retail pricing violations

    Back in 2017, Apple made a deal with New Jersey, and agreed to follow the state's visible pricing laws. Eight years later, in 2025, inspectors found widespread violations at Apple retail locations in the state, prompting Apple to revise its in-store pricing practices.

    So, the New Jersey Office of the Attorney General says that Apple will pay a $150,000 penalty because of the repeated inspection failures. The settlement addresses allegations that Apple failed to display prices clearly in New Jersey stores for years, despite the 2017 consent order.

    So, for Apple, this is a repeat issue, not a first offense.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Gemini, privacy, and the new Apple Creator Studio on the AppleInsider Podcast Fri, 16 Jan 2026 14:51:43 +0000
    They weren't unexpected, but they were still surprises — and they both leave us with so many questions: Google Gemini is coming to Apple Intelligence, and Apple Creator Studio is bring brilliant apps at a price.

    Colorful glowing diamond with rainbow gradient centered on black background, surrounded by looping neon ribbons, and a small black circle in the corner containing lowercase letters ai
    Google Gemini is coming to Apple Intelligence

    There's actually been two different Google Gemini announcements this week, and it's almost as if Google is trying to piggyback on the one about Apple. For where the Apple and Google deal ensures user privacy, the separate Gemini announcement truly does not.

    Nonetheless, it looks like Google Gemini is going to become part of our lives now — although it's also looking like we won't hear Apple speak its name ever again. Here's what's really going on, at least as far as the slow teasing out of news makes it look like today.


    Continue Reading on AppleInsider | Discuss on our Forums
  • MacBook Pro OLED screens may have already started production Fri, 16 Jan 2026 14:03:42 +0000
    A new rumor claims that mass production has begun early for an OLED display that could be intended for a future MacBook Pro.

    A MacBook Pro without an OLED display.
    A MacBook Pro without an OLED display.

    Apple has been expected to add a OLED screen to the MacBook Pro, with the most recent rumors predicting a late 2026 release. A new report raises the possibility that Apple may release its OLED MacBook Pro early.

    According to leaker yeux1122, Samsung Display has begun mass production of what's described as the 8.6 generation OLED screens. The leak claims that Samsung was originally expected to produce these screens from the second quarter of 2026.


    Rumor Score: 🤯 Likely


    Continue Reading on AppleInsider | Discuss on our Forums
  • Nvidia is squeezing out Apple for TSMC foundry capacity Fri, 16 Jan 2026 14:26:20 +0000
    Apple will still benefit as TMSC confirms large-volume production has started for its newest 2nm processor, but as Nvidia ascends, Apple's days of being able to command the firm's entire output are over — for now.

    Colorful, close-up view of silicon wafers with intricate grid patterns and reflective iridescent surfaces, displaying a spectrum of colors under light.
    Dies on wafers - Image Credit: TSMC

    Back in 2020, Apple was able to buy TSMC's entire 3-nanometer production capacity and use it in the iPhone, iPad, and Mac. Not only did that mean Apple had the then fastest processors, but no one else could buy them.

    According to semiconductor journalist Tim Culpan, however, that may have been the last time Apple was able to do this. That's because TSMC's production of AI processors for Nvidia and AMD is taking more of the firm's capacity.


    Continue Reading on AppleInsider | Discuss on our Forums
  • From charming to consumer corporation -- how Apple has branded itself over the years. Fri, 16 Jan 2026 13:09:54 +0000
    Apple began with Steve Jobs and Steve Wozniak marketing primarily at trade shows, and now the iPhone has made it $4 trillion firm with ubiquitous advertising. But throughout, Apple has used sometimes terrible branding to sell us all on its products.

    Man in a suit gesturing, casual-dressed man stands beside him; old Apple logo on left, colorful Apple logo on right.
    Iconic Apple marketing, L-R: original logo, "I'm a PC" campaign, and the famous six-colors logotype — images credit: Apple

    Today you know that every smart strapline like "thinpossible" or "Awe Dropping" has been user-tested, surgically examined by committees, and only put out when Apple is certain it's the greatest thing ever. That's far from unreasonable or even unusual, but it's a striking difference to how it used to be — and even to how the entire name of the company was chosen.

    This is a case of where you can pick from at least a handful of origin stories, usually involving Steve Jobs driving by an apple orchard. Or maybe living on one.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Inside Apple Podcasts: how to get the most out of Apple's player Thu, 10 Apr 2025 12:37:35 +0000
    Podcasts are wildly popular, and its easy to see why — from sports to true crime, comedy, news, and more, there seems to be one for nearly anything. Here's how to use Apple Podcasts and get the most out of your podcast experience.

    Apple Podcasts
    Apple Podcasts

    Unless you've been living under a rock, you've probably heard of podcasts by now. You might even be among the millions of people listening to them.

    Apple has anticipated this, which is why it includes its Apple Podcasts app on almost all its devices.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Analyst claims to know full Apple's iPhone 18 Pro & iPhone Fold specs Fri, 16 Jan 2026 12:20:20 +0000
    Analyst Jeff Pu has provided a short spec sheet for the iPhone 18 Pro, iPhone 18 Pro Max, and iPhone Fold due in September, which corroborates several previous rumors.

    A render of the iPhone Fold open partially exposing the hinge. The dual horizontal cameras on one side, a small display with selfie camera cutout on the other.
    iPhone Fold rumors are solidifying

    Rumors are beginning to solidify around Apple's fall iPhone launch, which may only feature three premium models. While Jeff Pu's data can sometimes be accurate, his assumptions and timelines tend to be too aggressive.

    The analyst has released a new note, which was detailed by 9to5Mac. The report itself isn't of much consequence, as it's the usual prediction of how Apple might perform in 2026, but it does include a spec table for the upcoming iPhone models.


    Rumor Score: 🤔 Possible


    Continue Reading on AppleInsider | Discuss on our Forums
  • There's a new Apple monitor imminent, but what it will be is anybody's guess Thu, 15 Jan 2026 19:02:03 +0000
    An unannounced Apple display has appeared in a Chinese regulatory database, offering the clearest sign yet that Apple is preparing to refresh one of its long-stagnant external monitors very soon.

    Dual widescreen monitors on a dark desk display a vivid ocean sunset with rocky coastline; a slim keyboard and mouse rest in front, creating a sleek modern workstation setup
    Apple's Studio Display

    The filing reveals a new model number and describes a high-performance LCD display, suggesting Apple may be updating the Studio Display, Pro Display XDR, or repositioning both. Though details are sparse, the timing and language imply the product is near release.

    The database listing is an Apple display with model number A3350, which matches no current product. Apple typically files regulatory paperwork late in development, generally months but sometimes weeks before announcements.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Spotify Premium jumps to $12.99/mo, $2 more expensive than Apple Music Thu, 15 Jan 2026 18:51:20 +0000
    Spotify subscribers — prepare for a price hike, as the company has announced another increase for users in the U.S., Estonia, and Latvia driving the cost notably above Apple Music.

    Spotify logo with black circle containing three curved sound-wave lines, next to bold black word Spotify on a plain white background
    Image credit: Spotify

    "Thank you for being a valued Premium subscriber," the announcement starts. "Starting on your billing date in February, your subscription price will change from $11.99/month to $12.99/month."

    The company claims the price hike is being done to "keep delivering a great experience."


    Continue Reading on AppleInsider | Discuss on our Forums
  • Apple chipmaker TSMC is spending big on next-gen silicon, AI chips, and the U.S. Thu, 15 Jan 2026 17:40:05 +0000
    Main Apple chip partner TSMC had a great earnings report, and to keep up with demand, it is increasing its manufacturing spend by almost 40%. It's not clear how much this will benefit Apple.

    TSMC's headquarters in Hsinchu, Taiwan
    TSMC's headquarters in Hsinchu, Taiwan

    The news comes after TMSC announced bumper quarterly results, which saw its profits increase by 35%. The company's strong performance is mostly thanks to the current AI boom, which has seen other chipmakers reap similar benefits.

    Both TSMC in an earnings report transcript and The Economic Times report that TSMC expects record revenue, thanks, in part, to its place at the forefront of chip technologies.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Disney Plus & Hulu bundle drops to $9.99 for a month, best streaming deal of 2026 Thu, 15 Jan 2026 17:06:00 +0000
    Disney is running a limited-time promotion offering one month of the Disney Plus and Hulu bundle for $9.99, a 26% drop from the standard pricing.

    Split screen showing Disney Plus logo on a blue gradient background on the left and Hulu logo in bright green text on a solid black background on the right
    Stream original programming with this Disney Plus and Hulu deals - Image credit: Disney, Hulu

    The $9.99 promotion is available to new and eligible returning subscribers, which Disney defines as users not currently subscribed to Disney Plus or any Disney bundle. The promotion applies exclusively to the ad-supported bundle tier, and auto-renews at $12.99/mo or the then-current regular monthly price unless canceled.

    Get 1 month of Disney+ and Hulu for $9.99


    Continue Reading on AppleInsider | Discuss on our Forums
  • Verizon offers $20 account credit following daylong cellular outage Thu, 15 Jan 2026 16:41:20 +0000
    Verizon is preparing to offer an olive branch to its customers after an outage took its cellular service offline for nearly ten hours. Here's how to get it.

    Large illuminated Verizon logo on a dark wall, with bold white letters and a bright red checkmark at the end, likely displayed in a modern indoor exhibition space
    Verizon to issue $20 account credit after 10-hour outage

    On Wednesday, around 12:30 PM ET, Verizon faced an outage that impacted hundreds of thousands of customers. Customers reported no service, with iPhone owners reporting an "SOS" where their network icon should have been.

    While users could still make calls and send texts via Wi-Fi, those away from an access point were left without service for up to ten hours.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Setapp Mobile EU app store, CleanMyMac Business both close down for good Thu, 15 Jan 2026 15:23:11 +0000
    MacPaw's third-party iOS App Store has failed in the EU, and at the same time, the company's mobile device management service is shutting down too.

    Three smartphones displaying an app, next to a large icon with a grid of six beige diamonds on a dark background with rounded corners.
    Setapp Mobile is closing down

    MacPaw's Setapp, which is continuing, is a single subscription for a suite of major Mac apps, but the company attempted to launch an EU-only iOS version in 2024. That was expressly to take advantage of how the European Union has forced Apple to allow rival stores on iOS.

    Now in a statement on its support pages, MacPaw says that this Setapp Mobile service is being shut down on February 16, 2026. It's "due to still-evolving and complex business term."


    Continue Reading on AppleInsider | Discuss on our Forums
  • Unsurprisingly, iPhone 17e will get Dynamic Island, keep 60Hz screen Thu, 15 Jan 2026 12:28:38 +0000
    The latest rumor regarding the forthcoming iPhone 17e is that despite an improved processor, the only difference coming to the display when compared to the iPhone 16e is a Dynamic Island.

    A smartphone, notebook, and pen are placed on a desk next to a laptop.
    iPhone 16e

    Having most recently claimed that there will be few visible changes to the iPhone 18 range, leaker Digital Chat Station is now saying the same about the iPhone 17e. Specifically, in a new leak on Chinese social media platform Weibo, he or she says that the iPhone 17e will keep the same screen as on the iPhone 16e.

    That means that it will be a 6.1-inch OLED display running at 60Hz. Which further means that the iPhone 17e will not get ProMotion, so it won't have an always-on display.


    Rumor Score: 🤔 Possible


    Continue Reading on AppleInsider | Discuss on our Forums
  • Grok now bans illegal porn generation, after monetizing it Thu, 15 Jan 2026 14:04:14 +0000
    Following worldwide pressure including bans in certain countries, Elon Musk's X has announced it will no longer allow Grok to create child porn and deepfake nudes. There is a catch.

    iPhone App Store screen showing Grok app listing with black logo and silver circular symbol, tagline Worlds Smartest AI Advisor, and a blue Get button with in app purchases text
    Grok on the App Store

    Users of Musk's AI tool Grok has been taking real-life photographs of women and children, and having it generate images that are pornographic. Musk originally described it as "way funnier," then later insisted that it was simply never happening — after he made it a paid premium feature.

    When we were looking into it, it took less than 10 seconds to find illegal content on X, generated by Grok. After Musk monetized the illegal porn generation, it took about 30 seconds, so it wasn't that effective a countermeasure.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Inside iMovie - How to get started with video editing on iPhone and iPad Mon, 29 Sep 2025 10:59:23 +0000
    Strive to become the next Steven Spielberg — or at least make your holiday videos better to watch — by using iMovie. Here's how to use it on your iPhone and iPad.

    Smartphone and tablet displaying a video editing app featuring the Eiffel Tower at night with different filters and options visible.
    Inside iMovie

    When it comes to video, most iPhone users will record using the Camera app and leave what's recorded in Photos as-is. If they're using social media a lot, they may just record the footage in there and use whatever tools are built into the app to create something entertaining.

    For those who care more about what their video says or looks like, they may start to look at dedicated editing tools. Yes, you can make a Memory in Photos using videos and can trim clips down to size, but you're not constructing a proper story that's under your direct control.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Apple AI research shows how MLLMs understand, generate, search for images Wed, 14 Jan 2026 23:58:07 +0000
    Apple's researchers continue to focus on multimodal LLMs, with studies exploring their use for image generation, understanding, and multi-turn web searches with cropped images.

    Smartphone standing on a desk displaying a colorful cartoon puppy sitting beside a latte, with avatar customization options and suggestion icons beneath, against a blurred purple and blue background
    Apple's AI studies explore the use of multimodal LLMs with images.

    With iOS 18, Apple made it possible to generate images on an iPhone through local AI models. Image Playground lets you create cartoon-like photos of just about anything, all without a Wi-Fi connection.

    Now, the company is continuing its image-related endeavors through research that explores how multi-modal LLMs use, generate, and understand images.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Level up your manicure with the Apple Watch charger Wed, 14 Jan 2026 22:14:43 +0000
    The nail polish girlies have found a new favorite tool to use with magnetic cat eye nail polish — the Apple Watch charger.

    Close-up of short square nails painted glossy black with a shimmering blue magnetic gradient, creating a galaxy-like effect under bright light on a finger and full hand.
    The method and the madness | Image credit: Reddit user peninkling

    If you keep up with nail polish trends — and I'm assuming many of you don't — magnetic nail polish is having a big moment right now. It's interesting, because magnetic polish is hardly new. I remember seeing it hit the shelves of bougier shops in my early 20s, only to trickle down to drugstore brands a year or two later.

    What is new, however, are the inventive ways people are using it. Just about anything that has its own localized magnetic field has been used to alter magnetic polish, and now that growing list includes the Apple Watch charger.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Apple Creator Studio takes one more step to an app subscription future Wed, 14 Jan 2026 21:23:32 +0000
    The Apple Creator Studio doesn't quite mean the end of perpetual software licenses, but in a new interview, it's clear that Apple expects consumers to embrace paying for bundles monthly instead of once.

    Person with headphones works at a wooden desk with dual monitors, laptop, microphone arm, studio light, and speakers, editing audio or video in a bright home office setup
    Apple Creator Studio is a compelling package for content creators - Image Credit: Apple

    The launch of the new Apple Creator Studio promised users access to media tools like Logic Pro and Final Cut Pro on Mac and iPad for a relatively low fee of $12.99 per month. However, at the same time, Apple is keeping the apps open for purchase with perpetual licenses at the same time.

    In an interview with CineD on Wednesday, Bryan O'Neil Hughes and John Danty of Apple's product marketing offered some clarification on what the bundle ultimately means for creative professionals who take up Apple's bundle.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Compared: Apple Creator Studio vs Adobe Creative Cloud Pro Wed, 14 Jan 2026 19:39:51 +0000
    Apple Creator Studio is taking on Adobe's dominant Creative Cloud by offering a selection of apps to create content, paid monthly. There are massive differences between the two subscription services, and some nuance to picking a package. Here's how they compare, and what you can do to fill the gaps.

    Colorful Adobe Creative Cloud logo centered on black background, surrounded by various neon-style app icons for drawing, audio, presentations, charts, and design tools arranged in an arc.
    Apple Creator Studio vs Adobe Creative Cloud Pro

    On Tuesday, Apple decided that it wanted to take on the ten-ton gorilla of creative apps, Adobe, on its own turf. The Apple Creator Studio is a collection of apps for editing video, creating music, producing art, and other creativity tasks, all within Apple's hardware ecosystem.

    At a high-level view, that's precisely what Adobe Creative Cloud Pro has provided to professionals for many years, mostly under its previous non-Pro form — a bunch of tools for subscribers to make practically any digital media they want, to a high level.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Verizon wireless outage left iPhones stuck in SOS across the U.S. Thu, 15 Jan 2026 12:29:56 +0000
    Verizon customers across the U.S. experienced a wireless network outage on Wednesday that disrupted calls, texts, and mobile data.

    Large 3D white Verizon logo with a red checkmark at the end, centered on a dark, industrial warehouse background with overhead lights reflecting on the floor
    Verizon outage is affecting people across the United States

    Reports of an outage began surfacing earlier Wednesday and spread quickly across multiple regions. Verizon acknowledged service disruptions and said restoration work was underway, but the company didn't shared a timeline beyond "work through the night."

    Account credits will be provided to affected customers.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Don't give Gemini your personal data, wait for Apple Intelligence-powered Siri Wed, 14 Jan 2026 17:48:44 +0000
    Google Gemini is now asking users to let it train on their personal email and photos, and offering up bizarre reasons why this is meant to be good. It is not, and the timing of the news is suspicious.

    Dark background with a glowing blue and white star above the word Gemini, surrounded by wavy lines.
    Google Gemini

    So on the one hand, Google Gemini is now coming to Apple Intelligence, and it's coming with all of Apple's rigid privacy protections. But on the other, in the same week that was announced, Google is suddenly trying to get Gemini users to entirely ignore privacy.

    Specifically, Google wants Gemini users in the US to join a beta program and authorize the service to scrape Gmail, Google Photos, and YouTube. Google calls this Personal Intelligence, and the company's Josh Woodward offers real-world examples of how great this is.


    Continue Reading on AppleInsider | Discuss on our Forums
  • iPhone Fold's titanium & Liquid Metal rumors just won't go away Wed, 14 Jan 2026 16:45:54 +0000
    A leaker has reiterated claims about the iPhone Fold using a liquid metal hinge and titanium body, going against the current belief that it will have an aluminum body.

    Close-up of a sleek gold foldable smartphone half-open, highlighting dual rear cameras, a small flash, and reflective inner screen with soft warm background.
    A render of what the iPhone Fold could look like - Image Credit: AppleInsider

    Unlike a regular static iPhone, the iPhone Fold is designed to be folded in half and opened up on a regular basis. This potential point of weakness means Apple has to be extremely careful about the types of material that it uses for the frequently-rumored model.

    According to claims made by leaker Yeux1122 in a Naver post on Wednesday, Apple's choice of materials for the casing and hinge of the iPhone Fold will be titanium and liquid metal, respectively. While the claims about the materials aren't new, the leaker adds that they are better than in previous versions.


    Rumor Score: 🤔 Possible


    Continue Reading on AppleInsider | Discuss on our Forums
  • 'Civilization VII Arcade Edition' coming to Apple Arcade with big caveats Wed, 14 Jan 2026 15:52:00 +0000
    "Civilization VII Arcade Edition" is coming to Apple Arcade on February 5, bringing the long-running strategy series to more Apple platforms. But, there are some big limitations.

    Elegant marble and gold decorative arches with small world landmarks surround centered text reading Sid Meier's Civilization VII Arcade Edition on a soft, light stone-textured background
    "Civilization VII Arcade Edition" is coming to Apple Arcade

    The game will be available on iPhone, iPad, and Mac through an Apple Arcade subscription, without ads or in-app purchases. For Apple users, that removes the usual upfront cost tied to a franchise long sold as a premium PC release.

    Civilization VII Arcade Edition delivers the core turn-based strategy experience, with players guiding civilizations through distinct historical ages shaped by long-term choices. The release is positioned as a full entry adapted for Apple devices, not a simplified mobile offshoot.


    Continue Reading on AppleInsider | Discuss on our Forums
  • Apple Vision Pro owners will get a great assortment of classic arcade games in VR soon Wed, 14 Jan 2026 15:54:49 +0000
    In February, Apple Arcade is getting the nostalgic "Retrocade" for Apple Vision Pro, allowing players to walk around, play classic arcade titles, and inspect cabinets in VR.

    Yellow Pac-Man arcade cabinet in a cozy living room, beside a gray sofa with colorful pillows and a white shelving unit holding plants, books, and a typewriter
    A Pac-Man arcade cabinet in Retrocade on an Apple Vision Pro - Image Credit: Apple

    Apple Arcade adds more games to its catalog each month, but sometimes it turns into a history lesson. For February, it will be an opportunity for gamers to relive some of the best arcade games from the 1980's.

    Retrocade from Resolution Games will be added to Apple Arcade on February 5. As the name implies, it's an app that aims to give the arcade experience to a modern audience, by including a selection of classic titles to play.


    Continue Reading on AppleInsider | Discuss on our Forums
  • iPhone 18 design won't change much, except for Dynamic Island Wed, 14 Jan 2026 15:28:46 +0000
    The iPhone 18 Pro and iPhone 18 Pro Max are claimed to have new under-display technology, but it's not known just how much that will reduce the Dynamic Island.

    Silver smartphone with a matte wrapper, triple camera setup on textured surface, and visible side buttons.
    The iPhone 18 Pro is expected to deliver a multitude of improvements.

    Previous leaks regarding the iPhone 18 range have said that the models will closely resemble their iPhone 17 counterparts. Now Digital Chat Station claims that most of the iPhone 18 range will keep to the same screen sizes as before, specifically:



Ars Technica



VentureBeat

  • Listen Labs raises $69M after viral billboard hiring stunt to scale AI customer interviews Fri, 16 Jan 2026 14:01:00 GMT

    Alfred Wahlforss was running out of options. His startup, Listen Labs, needed to hire over 100 engineers, but competing against Mark Zuckerberg's $100 million offers seemed impossible. So he spent $5,000 — a fifth of his marketing budget — on a billboard in San Francisco displaying what looked like gibberish: five strings of random numbers.

    The numbers were actually AI tokens. Decoded, they led to a coding challenge: build an algorithm to act as a digital bouncer at Berghain, the Berlin nightclub famous for rejecting nearly everyone at the door. Within days, thousands attempted the puzzle. 430 cracked it. Some got hired. The winner flew to Berlin, all expenses paid.

    That unconventional approach has now attracted $69 million in Series B funding, led by Ribbit Capital with participation from Evantic and existing investors Sequoia Capital, Conviction, and Pear VC. The round values Listen Labs at $500 million and brings its total capital to $100 million. In nine months since launch, the company has grown annualized revenue by 15x to eight figures and conducted over one million AI-powered interviews.

    "When you obsess over customers, everything else follows," Wahlforss said in an interview with VentureBeat. "Teams that use Listen bring the customer into every decision, from marketing to product, and when the customer is delighted, everyone is."

    Why traditional market research is broken, and what Listen Labs is building to fix it

    Listen's AI researcher finds participants, conducts in-depth interviews, and delivers actionable insights in hours, not weeks. The platform replaces the traditional choice between quantitative surveys — which provide statistical precision but miss nuance—and qualitative interviews, which deliver depth but cannot scale.

    Wahlforss explained the limitation of existing approaches: "Essentially surveys give you false precision because people end up answering the same question... You can't get the outliers. People are actually not honest on surveys." The alternative, one-on-one human interviews, "gives you a lot of depth. You can ask follow up questions. You can kind of double check if they actually know what they're talking about. And the problem is you can't scale that."

    The platform works in four steps: users create a study with AI assistance, Listen recruits participants from its global network of 30 million people, an AI moderator conducts in-depth interviews with follow-up questions, and results are packaged into executive-ready reports including key themes, highlight reels, and slide decks.

    What distinguishes Listen's approach is its use of open-ended video conversations rather than multiple-choice forms. "In a survey, you can kind of guess what you should answer, and you have four options," Wahlforss said. "Oh, they probably want me to buy high income. Let me click on that button versus an open ended response. It just generates much more honesty."

    The dirty secret of the $140 billion market research industry: rampant fraud

    Listen finds and qualifies the right participants in its global network of 30 million people. But building that panel required confronting what Wahlforss called "one of the most shocking things that we've learned when we entered this industry"—rampant fraud.

    "Essentially, there's a financial transaction involved, which means there will be bad players," he explained. "We actually had some of the largest companies, some of them have billions in revenue, send us people who claim to be kind of enterprise buyers to our platform and our system immediately detected, like, fraud, fraud, fraud, fraud, fraud."

    The company built what it calls a "quality guard" that cross-references LinkedIn profiles with video responses to verify identity, checks consistency across how participants answer questions, and flags suspicious patterns. The result, according to Wahlforss: "People talk three times more. They're much more honest when they talk about sensitive topics like politics and mental health."

    Emeritus, an online education company that uses Listen, reported that approximately 20% of survey responses previously fell into the fraudulent or low-quality category. With Listen, they reduced this to almost zero. "We did not have to replace any responses because of fraud or gibberish information," said Gabrielli Tiburi, Assistant Manager of Customer Insights at Emeritus.

    How Microsoft, Sweetgreen, and Chubbies are using AI interviews to build better products

    The speed advantage has proven central to Listen's pitch. Traditional customer research at Microsoft could take four to six weeks to generate insights. "By the time we get to them, either the decision has been made or we lose out on the opportunity to actually influence it," said Romani Patel, Senior Research Manager at Microsoft.

    With Listen, Microsoft can now get insights in days, and in many cases, within hours.

    The platform has already powered several high-profile initiatives. Microsoft used Listen Labs to collect global customer stories for its 50th anniversary celebration. "We wanted users to share how Copilot is empowering them to bring their best self forward," Patel said, "and we were able to collect those user video stories within a day." Traditionally, that kind of work would have taken six to eight weeks.

    Simple Modern, an Oklahoma-based drinkware company, used Listen to test a new product concept. The process took about an hour to write questions, an hour to launch the study, and 2.5 hours to receive feedback from 120 people across the country. "We went from 'Should we even have this product?' to 'How should we launch it?'" said Chris Hoyle, the company's Chief Marketing Officer.

    Chubbies, the shorts brand, achieved a 24x increase in youth research participation—growing from 5 to 120 participants — by using Listen to overcome the scheduling challenges of traditional focus groups with children. "There's school, sports, dinner, and homework," explained Lauren Neville, Director of Insights and Innovation. "I had to find a way to hear from them that fit into their schedules."

    The company also discovered product issues through AI interviews that might have gone undetected otherwise. Wahlforss described how the AI "through conversations, realized there were like issues with the the kids short line, and decided to, like, interview hundreds of kids. And I understand that there were issues in the liner of the shorts and that they were, like, scratchy, quote, unquote, according to the people interviewed." The redesigned product became "a blockbuster hit."

    The Jevons paradox explains why cheaper research creates more demand, not less

    Listen Labs is entering a massive but fragmented market. Wahlforss cited research from Andreessen Horowitz estimating the market research industry at roughly $140 billion annually, populated by legacy players — some with more than a billion dollars in revenue — that he believes are vulnerable to disruption.

    "There are very much existing budget lines that we are replacing," Wahlforss said. "Why we're replacing them is that one, they're super costly. Two, they're kind of stuck in this old paradigm of choosing between a survey or interview, and they also take months to work with."

    But the more intriguing dynamic may be that AI-powered research doesn't just replace existing spending — it creates new demand. Wahlforss invoked the Jevons paradox, an economic principle that occurs when technological advancements make a resource more efficient to use, but increased efficiency leads to increased overall consumption rather than decreased consumption.

    "What I've noticed is that as something gets cheaper, you don't need less of it. You want more of it," Wahlforss explained. "There's infinite demand for customer understanding. So the researchers on the team can do an order of magnitude more research, and also other people who weren't researchers before can now do that as part of their job."

    Inside the elite engineering team that built Listen Labs before they had a working toilet

    Listen Labs traces its origins to a consumer app that Wahlforss and his co-founder built after meeting at Harvard. "We built this consumer app that got 20,000 downloads in one day," Wahlforss recalled. "We had all these users, and we were thinking like, okay, what can we do to get to know them better? And we built this prototype of what Listen is today."

    The founding team brings an unusual pedigree. Wahlforss's co-founder "was the national champion in competitive programming in Germany, and he worked at Tesla Autopilot." The company claims that 30% of its engineering team are medalists from the International Olympiad in Informatics — the same competition that produced the founders of Cognition, the AI coding startup.

    The Berghain billboard stunt generated approximately 5 million views across social media, according to Wahlforss. It reflected the intensity of the talent war in the Bay Area.

    "We had to do these things because some of our, like early employees, joined the company before we had a working toilet," he said. "But now we fixed that situation."

    The company grew from 5 to 40 employees in 2024 and plans to reach 150 this year. It hires engineers for non-engineering roles across marketing, growth, and operations — a bet that in the AI era, technical fluency matters everywhere.

    Synthetic customers and automated decisions: what Listen Labs is building next

    Wahlforss outlined an ambitious product roadmap that pushes into more speculative territory. The company is building "the ability to simulate your customers, so you can take all of those interviews we've done, and then extrapolate based on that and create synthetic users or simulated user voices."

    Beyond simulation, Listen aims to enable automated action based on research findings. "Can you not just make recommendations, but also create spawn agents to either change things in code or some customer churns? Can you give them a discount and try to bring them back?"

    Wahlforss acknowledged the ethical implications. "Obviously, as you said, there's kind of ethical concerns there. Of like, automated decision making overall can be bad, but we will have considerable guardrails to make sure that the companies are always in the loop."

    The company already handles sensitive data with care. "We don't train on any of the data," Wahlforss said. "We will also scrub any sensitive PII automatically so the model can detect that. And there are times when, for example, you work with investors, where if you accidentally mention something that could be material, non public information, the AI can actually detect that and remove any information like that."

    How AI could reshape the future of product development

    Perhaps the most provocative implication of Listen's model is how it could reshape product development itself. Wahlforss described a customer — an Australian startup — that has adopted what amounts to a continuous feedback loop.

    "They're based in Australia, so they're coding during the day, and then in their night, they're releasing a Listen study with an American audience. Listen validates whatever they built during the day, and they get feedback on that. They can then plug that feedback directly into coding tools like Claude Code and iterate."

    The vision extends Y Combinator's famous dictum — "write code, talk to users" — into an automated cycle. "Write code is now getting automated. And I think like talk to users will be as well, and you'll have this kind of infinite loop where you can start to ship this truly amazing product, almost kind of autonomously."

    Whether that vision materializes depends on factors beyond Listen's control — the continued improvement of AI models, enterprise willingness to trust automated research, and whether speed truly correlates with better products. A 2024 MIT study found that 95% of AI pilots fail to move into production, a statistic Wahlforss cited as the reason he emphasizes quality over demos.

    "I'm constantly have to emphasize like, let's make sure the quality is there and the details are right," he said.

    But the company's growth suggests appetite for the experiment. Microsoft's Patel said Listen has "removed the drudgery of research and brought the fun and joy back into my work." Chubbies is now pushing its founder to give everyone in the company a login. Sling Money, a stablecoin payments startup, can create a survey in ten minutes and receive results the same day.

    "It's a total game changer," said Ali Romero, Sling Money's marketing manager.

    Wahlforss has a different phrase for what he's building. When asked about the tension between speed and rigor — the long-held belief that moving fast means cutting corners — he cited Nat Friedman, the former GitHub CEO and Listen investor, who keeps a list of one-liners on his website.

    One of them: "Slow is fake."

    It's an aggressive claim for an industry built on methodological caution. But Listen Labs is betting that in the AI era, the companies that listen fastest will be the ones that win. The only question is whether customers will talk back.

  • Kilo launches AI-powered Slack bot that ships code from a chat message Fri, 16 Jan 2026 14:00:00 GMT

    Kilo Code, the open-source AI coding startup backed by GitLab cofounder Sid Sijbrandij, is launching a Slack integration that allows software engineering teams to execute code changes, debug issues, and push pull requests directly from their team chat — without opening an IDE or switching applications.

    The product, called Kilo for Slack, arrives as the AI-assisted coding market heats up with multibillion-dollar acquisitions and funding rounds. But rather than building another siloed coding assistant, Kilo is making a calculated bet: that the future of AI development tools lies not in locking engineers into a single interface, but in embedding AI capabilities into the fragmented workflows where decisions actually happen.

    "Engineering teams don't make decisions in IDE sidebars. They make them in Slack," Scott Breitenother, Kilo Code's co-founder and CEO, said in an interview with VentureBeat. "The Slackbot allows you to do all this — and more — without leaving Slack."

    The launch also marks a partnership with MiniMax, the Hong Kong-based AI company that recently completed a successful initial public offering. MiniMax's M2.1 model will serve as the default model powering Kilo for Slack — a decision the company frames as a statement about the closing gap between open-weight and proprietary frontier models.

    How Kilo for Slack turns team conversations into pull requests without leaving the chat

    The integration operates on a simple premise: Slack threads often contain the context needed to fix a bug or implement a feature, but that context gets lost the moment a developer switches to their code editor.

    With Kilo for Slack, users mention @Kilo in a Slack thread, and the bot reads the entire conversation, accesses connected GitHub repositories, and either answers questions about the codebase or creates a branch and submits a pull request.

    A typical interaction might look like this: A product manager reports a bug in a Slack channel. Engineers discuss potential causes. Instead of someone copying the conversation into their IDE and re-explaining the problem to an AI assistant, a developer simply types: "@Kilo based on this thread, can you implement the fix for the null pointer exception in the Authentication service?"

    The bot then spins up a cloud agent, reads the thread context, implements the fix, and pushes a pull request — all visible in Slack.

    The company says the entire process eliminates the need to copy information between apps or jump between windows — developers can trigger complex code changes with nothing more than a single message in Slack.

    Why Kilo says Cursor and Claude Code fall short when developers need multi-repo context

    Kilo's launch explicitly positions the product against two leading AI coding tools: Cursor, which raised $2.3 billion at a $29.3 billion valuation in November, and Claude Code, Anthropic's agentic coding tool.

    Breitenother outlined specific limitations he sees in both products' Slack capabilities.

    "The Cursor Slack integration is configured on a single-repository basis per workspace or channel," he said. "As a result, if a Slack thread references multiple repositories, users need to manually switch or reconfigure the integration to pull in that additional context."

    On Anthropic's offering, he added: "Claude Code documentation for Slack shows how Claude can be added to a workspace and respond to mentions using the surrounding conversation context. However, it does not describe persistent, multi-turn thread state or task-level continuity across longer workflows. Each interaction is handled based on the context included at the time of the prompt, rather than maintaining an evolving execution state over time."

    Kilo claims its integration works across multiple repositories simultaneously, maintains conversational context across extended Slack threads, and enables handoffs between Slack, IDEs, cloud agents, and the command-line interface.

    Kilo picks a Chinese AI company's model as its default—and addresses enterprise security concerns head-on

    Perhaps the most provocative element of the announcement is Kilo's choice of default model. MiniMax is headquartered in Shanghai and recently went public in Hong Kong — a lineage that may raise eyebrows among enterprise customers wary of sending proprietary code through Chinese infrastructure.

    Breitenother addressed the concern directly: "MiniMax's recent Hong Kong IPO drew backing from major global institutional investors, including Baillie Gifford, ADIA, GIC, Mirae Asset, Aspex, and EastSpring. This speaks to strong global confidence in models built for global users."

    He emphasized that MiniMax models are hosted by major U.S.-compliant cloud providers. "MiniMax M2-series are global leading open-source models, and are hosted by many U.S. compliant cloud providers such as AWS Bedrock, Google Vertex and Microsoft AI Foundry," he said. "In fact, MiniMax models were featured by Matt Garman, the AWS CEO, during this year's re:Invent keynote, showing they're ready for enterprise use at scale."

    The company stresses that Kilo for Slack is fundamentally model-agnostic. "Kilo doesn't force customers into any single model," Breitenother said. "Enterprise customers choose which models they use, where they're hosted, and what fits their security, compliance, and risk requirements. Kilo offers access to more than 500 models, so teams can always choose the right model for the job."

    The decision to default to M2.1 reflects Kilo's broader thesis about the AI market. According to the company, the performance gap between open-weight and proprietary models has narrowed from 8 percent to 1.7 percent on several key benchmarks. Breitenother clarified that this figure "refers to convergence between open and closed models as measured by the Stanford AI Index using major general benchmarks like HumanEval, MATH, and MMLU, not to any specific agentic coding evaluation."

    In third-party evaluations, M2.1 has performed competitively. "In LMArena, an open platform for community-driven AI benchmarking, M2.1 achieved a number-four ranking, right after OpenAI, Anthropic, and Google," Breitenother noted. "What this shows is that M2.1 competes with frontier models in real-world coding workflows, as judged directly by developers."

    What happens to your code when you @mention an AI bot in Slack

    For engineering teams evaluating the tool, a critical question is what happens to sensitive code and conversations when routed through the integration.

    Breitenother walked through the data flow: "When someone mentions @Kilo in Slack, Kilo reads only the content of the Slack thread where it's mentioned, along with basic metadata needed to understand context. It does not have blanket access to a workspace. Access is governed by Slack's standard permission model and the scopes the customer approves during installation."

    For repository access, he added: "If the request requires code context, Kilo accesses only the GitHub repositories the customer has explicitly connected. It does not index unrelated repos. Permissions mirror the access level granted through GitHub, and Kilo can't see anything the user or workspace hasn't authorized."

    The company states that data is not used to train models and that output visibility follows existing Slack and GitHub permissions.

    A particularly thorny question for any AI system that can push code directly to repositories is security. What prevents an AI-generated vulnerability from being merged into production?

    "Nothing gets merged automatically," Breitenother said. "When the Kilo Slackbot opens a pull request from a Slack thread, it follows the same guardrails teams already rely on today. The PR goes through existing review workflows and approval processes before anything reaches production."

    He added that Kilo can automatically run its built-in code review feature on AI-generated pull requests, "flagging potential issues or security concerns before it ever reaches a developer for review."

    The open-source paradox: why Kilo believes giving away its code won't kill the business

    Kilo Code sits in an increasingly common but still tricky position: the open-source company charging for hosted services. The complete IDE extension is open-source under an Apache 2.0 license, but Kilo for Slack is a paid, hosted product.

    The obvious question: What stops a well-funded competitor — or even a customer — from forking the code and building their own version?

    "Forking the code isn't what worries us, because the code itself isn't the hardest part," Breitenother said. "A competitor could fork the repository tomorrow. What they wouldn't get is the infrastructure that safely executes agentic workflows across Slack, GitHub, IDEs, and cloud agents. The experience we've built operating this at scale across many teams and repositories. The trust, integrations, and enterprise-ready controls customers expect out of the box."

    He drew parallels to other successful open-source companies: "Open core drives adoption and trust, while the hosted product delivers convenience, reliability, and ongoing innovation. Customers aren't paying for access to code. They're paying for a system that works every day, securely, at scale."

    Inside the $29 billion "vibe coding" market that Kilo wants to disrupt

    Kilo enters a market that has attracted extraordinary attention and capital over the past year. The practice of using large language models to write and modify code — popularly known as "vibe coding," a term coined by OpenAI co-founder Andrej Karpathy in February 2025 — has become a central focus of enterprise AI investment.

    Microsoft CEO Satya Nadella disclosed in April that AI-generated code now accounts for 30 percent of Microsoft's codebase. Google acquired senior employees from AI coding startup Windsurf in a $2.4 billion transaction in July. Cursor's November funding round valued the company at $29.3 billion.

    Kilo raised $8 million in seed funding in December 2025 from Breakers, Cota Capital, General Catalyst, Quiet Capital, and Tokyo Black. Sijbrandij, who stepped down as GitLab CEO in 2024 to focus on cancer treatment but remains board chair, contributed early capital and remains involved in day-to-day strategy.

    Asked about non-compete considerations given GitLab's own AI investments, Breitenother was brief: "There are no non-compete issues. Kilo is building a fundamentally different approach to AI coding."

    Notably, GitLab disclosed in a recent SEC filing that it paid Kilo $1,000 in exchange for a right of first refusal for 10 business days should the startup receive an acquisition proposal before August 2026.

    When asked to name an enterprise customer using the Slack integration in production, Breitenother declined: "That's not something we can disclose."

    How a 34-person startup plans to outmaneuver OpenAI and Anthropic in AI coding

    The most significant threat to Kilo's position may come not from other startups but from the frontier AI labs themselves. OpenAI and Anthropic are both building deeper integrations for coding workflows, and both have vastly greater resources.

    Breitenother argued that Kilo's advantage lies in its architecture, not its model performance.

    "We don't think the long-term moat in AI coding is raw compute or who ships a Slack agent first," he said. "OpenAI and Anthropic are world-class model companies, and they'll continue to build impressive capabilities. But Kilo is built around a different thesis: the hard problem isn't generating code, it's integrating AI into real engineering workflows across tools, repos, and environments."

    He outlined three areas where he believes Kilo can differentiate:

    "Workflow depth: Kilo is designed to operate across Slack, IDEs, cloud agents, GitHub, and the CLI, with persistent context and execution. Even with OpenAI or Anthropic Slack-native agents, those agents are still fundamentally model-centric. Kilo is workflow-centric."

    "Model flexibility: We're model-agnostic by design. Teams don't have to bet on one frontier model or vendor roadmap. That's difficult for companies like OpenAI or Anthropic, whose incentives are naturally aligned with driving usage toward their own models first."

    "Platform neutrality: Kilo isn't trying to pull developers into a closed ecosystem. It fits into the tools teams already use."

    The future of AI-assisted software development may belong to whoever solves the integration problem first

    Kilo's launch reflects a maturing phase in the AI coding market. The initial wave of tools focused on proving that large language models could generate useful code. The current wave is about integration — fitting AI capabilities into the messy reality of how software actually gets built.

    That reality involves context fragmented across Slack threads, GitHub issues, IDE windows, and command-line sessions. It involves teams that use different models for different tasks and organizations with complex compliance requirements around data residency and model providers.

    Kilo is betting that the winners in this market will not be the companies with the best models, but those that best solve the integration problem — meeting developers in the tools they already use rather than forcing them into new ones.

    Kilo for Slack is available now for teams with Kilo Code accounts. Users connect their GitHub repositories through Kilo's integrations dashboard, add the Slack integration, and can then mention @Kilo in any channel where the bot has been added. Usage-based pricing matches the rates of whatever model the team selects.

    Whether a 34-person startup can execute on that vision against competitors with billions in capital remains an open question. But if Breitenother is right that the hard problem in AI coding isn't generating code but integrating into workflows, Kilo may have picked the right fight. After all, the best AI in the world doesn't matter much if developers have to leave the conversation to use it.

  • Claude Code just got updated with one of the most-requested user features Thu, 15 Jan 2026 19:37:00 GMT

    Anthropic's open source standard, the Model Context Protocol (MCP), released in late 2024, allows users to connect AI models and the agents atop them to external tools in a structured, reliable format. It is the engine behind Anthropic's hit AI agentic programming harness, Claude Code, allowing it to access numerous functions like web browsing and file creation immediately when asked.

    But there was one problem: Claude Code typically had to "read" the instruction manual for every single tool available, regardless of whether it was needed for the immediate task, using up the available context that could otherwise be filled with more information from the user's prompts or the agent's responses.

    At least until last night. The Claude Code team released an update that fundamentally alters this equation. Dubbed MCP Tool Search, the feature introduces "lazy loading" for AI tools, allowing agents to dynamically fetch tool definitions only when necessary.

    It is a shift that moves AI agents from a brute-force architecture to something resembling modern software engineering—and according to early data, it effectively solves the "bloat" problem that was threatening to stifle the ecosystem.

    The 'Startup Tax' on Agents

    To understand the significance of Tool Search, one must understand the friction of the previous system. The Model Context Protocol (MCP), released in 2024 by Anthropic as an open source standard was designed to be a universal standard for connecting AI models to data sources and tools—everything from GitHub repositories to local file systems.

    However, as the ecosystem grew, so did the "startup tax."

    Thariq Shihipar, a member of the technical staff at Anthropic, highlighted the scale of the problem in the announcement.

    "We've found that MCP servers may have up to 50+ tools," Shihipar wrote. "Users were documenting setups with 7+ servers consuming 67k+ tokens."

    In practical terms, this meant a developer using a robust set of tools might sacrifice 33% or more of their available context window limit of 200,000 tokens before they even typed a single character of a prompt, as AI newsletter author Aakash Gupta pointed out in a post on X.

    The model was effectively "reading" hundreds of pages of technical documentation for tools it might never use during that session.

    Community analysis provided even starker examples.

    Gupta further noted that a single Docker MCP server could consume 125,000 tokens just to define its 135 tools.

    "The old constraint forced a brutal tradeoff," he wrote. "Either limit your MCP servers to 2-3 core tools, or accept that half your context budget disappears before you start working."

    How Tool Search Works

    The solution Anthropic rolled out — which Shihipar called "one of our most-requested features on GitHub" — is elegant in its restraint. Instead of preloading every definition, Claude Code now monitors context usage.

    According to the release notes, the system automatically detects when tool descriptions would consume more than 10% of the available context.

    When that threshold is crossed, the system switches strategies. Instead of dumping raw documentation into the prompt, it loads a lightweight search index.

    When the user asks for a specific action—say, "deploy this container"—Claude Code doesn't scan a massive, pre-loaded list of 200 commands. Instead, it queries the index, finds the relevant tool definition, and pulls only that specific tool into the context.

    "Tool Search flips the architecture," Gupta analyzed. "The token savings are dramatic: from ~134k to ~5k in Anthropic’s internal testing. That’s an 85% reduction while maintaining full tool access."

    For developers maintaining MCP servers, this shifts the optimization strategy.

    Shihipar noted that the `server instructions` field in the MCP definition—previously a "nice to have"—is now critical. It acts as the metadata that helps Claude "know when to search for your tools, similar to skills."

    'Lazy Loading' and Accuracy Gains

    While the token savings are the headline metric—saving money and memory is always popular—the secondary effect of this update might be more important: focus.

    LLMs are notoriously sensitive to "distraction." When a model's context window is stuffed with thousands of lines of irrelevant tool definitions, its ability to reason decreases. It creates a "needle in a haystack" problem where the model struggles to differentiate between similar commands, such as `notification-send-user` versus `notification-send-channel`.

    Boris Cherny, Head of Claude Code, emphasized this in his reaction to the launch on X: "Every Claude Code user just got way more context, better instruction following, and the ability to plug in even more tools."

    The data backs this up. Internal benchmarks shared by the community indicate that enabling Tool Search improved the accuracy of the Opus 4 model on MCP evaluations from 49% to 74%.

    For the newer Opus 4.5, accuracy jumped from 79.5% to 88.1%.

    By removing the noise of hundreds of unused tools, the model can dedicate its "attention" mechanisms to the user's actual query and the relevant active tools.

    Maturing the Stack

    This update signals a maturation in how we treat AI infrastructure. In the early days of any software paradigm, brute force is common. But as systems scale, efficiency becomes the primary engineering challenge.

    Aakash Gupta drew a parallel to the evolution of Integrated Development Environments (IDEs) like VSCode or JetBrains. "The bottleneck wasn’t 'too many tools.'

    It was loading tool definitions like 2020-era static imports instead of 2024-era lazy loading," he wrote. "VSCode doesn’t load every extension at startup. JetBrains doesn’t inject every plugin’s docs into memory."

    By adopting "lazy loading"—a standard best practice in web and software development—Anthropic is acknowledging that AI agents are no longer just novelties; they are complex software platforms that require architectural discipline.

    Implications for the Ecosystem

    For the end user, this update is seamless: Claude Code simply feels "smarter" and retains more memory of the conversation. But for the developer ecosystem, it opens the floodgates.

    Previously, there was a "soft cap" on how capable an agent could be. Developers had to curate their toolsets carefully to avoid lobotomizing the model with excessive context. With Tool Search, that ceiling is effectively removed. An agent can theoretically have access to thousands of tools—database connectors, cloud deployment scripts, API wrappers, local file manipulators—without paying a penalty until those tools are actually touched.

    It turns the "context economy" from a scarcity model into an access model. As Gupta summarized, "They’re not just optimizing context usage. They’re changing what ‘tool-rich agents’ can mean."

    The update is rolling out immediately for Claude Code users. For developers building MCP clients, Anthropic recommends implementing the `ToolSearchTool` to support this dynamic loading, ensuring that as the agentic future arrives, it doesn't run out of memory before it even says hello.

  • Why MongoDB thinks better retrieval — not bigger models — is the key to trustworthy enterprise AI Thu, 15 Jan 2026 18:00:00 GMT

    Agentic systems and enterprise search depend on strong data retrieval that works efficiently and accurately. Database provider MongoDB thinks its newest embeddings models help solve falling retrieval quality as more AI systems go into production.

    As agentic and RAG systems move into production, retrieval quality is emerging as a quiet failure point — one that can undermine accuracy, cost, and user trust even when models themselves perform well.

    The company launched four new versions of its embeddings and reranking models. Voyage 4 will be available in four modes: voyage-4 embedding, voyage-4-large, voyage-4-lite, and voyage-4-nano.  

    MongoDB said the voyage-4 embedding serves as its general-purpose model; MongoDB considers Voyage-4-large its flagship model. Voyage-4-lite focuses on tasks requiring little latency and lower costs, and voyage-4-nano is intended for more local development and testing environments or for on-device data retrieval. 

    Voyage-4-nano is also MongoDB’s first open-weight model. All models are available via an API and on MongoDB’s Atlas platform. 

    The company said the models outperform similar models from Google and Cohere on the RTEB benchmark. Hugging Face’s RTEB benchmark puts Voyage 4 as the top embedding model. 

    “Embedding models are one of those invisible choices that can really make or break AI experiences,” Frank Liu, product manager at MongoDB, said in a briefing. “You get them wrong, your search results will feel pretty random and shallow, but if you get them right, your application suddenly feels like it understands your users and your data.”

    He added that the goal of the Voyage 4 models is to improve the retrieval of real-world data, which often collapses once agentic and RAG pipelines go into production. 

    MongoDB also released a new multimodal embedding model, voyage-multimodal-3.5, that can handle documents that include text, images, and video. This model vectorizes the data and extracts semantic meaning from the tables, graphics, figures, and slides typically found in enterprise documents.

    Enterprise’s embeddings problems

    For enterprises, an agentic system is only as good as its ability to reliably retrieve the right information at the right time. This requirement becomes harder as workloads scale and context windows fragment.

    Several model providers target that layer of agentic AI. Google’s Gemini Embedding model topped the embedding leaderboards, and Cohere launched its Embed 4 multimodal model, which processes documents more than 200 pages long. Mistral said its coding-embedding model, Codestral Embedding, outperforms Cohere, Google, and even MongoDB’s Voyage Code 3. MongoDB argues that benchmark performance alone doesn’t address the operational complexity enterprises face in production.

    MongoDB said many clients have found that their data stacks cannot handle context-aware, retrieval-intensive workloads in production. The company said it's seeing more fragmentation with enterprises having to stitch together different solutions to connect databases with a retrieval or reranking model. To help customers who don’t want fragmented solutions, the company is offering its models through a single data platform, Atlas. 

    MongoDB’s bet is that retrieval can’t be treated as a loose collection of best-of-breed components anymore. For enterprise agents to work reliably at scale, embeddings, reranking, and the data layer need to operate as a tightly integrated system rather than a stitched-together stack.

  • Breaking through AI’s memory wall with token warehousing Thu, 15 Jan 2026 05:00:00 GMT

    As agentic AI moves from experiments to real production workloads, a quiet but serious infrastructure problem is coming into focus: memory. Not compute. Not models. Memory.

    Under the hood, today’s GPUs simply don’t have enough space to hold the Key-Value (KV) caches that modern, long-running AI agents depend on to maintain context. The result is a lot of invisible waste — GPUs redoing work they’ve already done, cloud costs climbing, and performance taking a hit. It’s a problem that’s already showing up in production environments, even if most people haven’t named it yet.

    At a recent stop on the VentureBeat AI Impact Series, WEKA CTO Shimon Ben-David joined VentureBeat CEO Matt Marshall to unpack the industry’s emerging “memory wall,” and why it’s becoming one of the biggest blockers to scaling truly stateful agentic AI — systems that can remember and build on context over time. The conversation didn’t just diagnose the issue; it laid out a new way to think about memory entirely, through an approach WEKA calls token warehousing.

    The GPU memory problem

    “When we're looking at the infrastructure of inferencing, it is not a GPU cycles challenge. It's mostly a GPU memory problem,” said Ben-David.

    The root of the issue comes down to how transformer models work. To generate responses, they rely on KV caches that store contextual information for every token in a conversation. The longer the context window, the more memory those caches consume, and it adds up fast. A single 100,000-token sequence can require roughly 40GB of GPU memory, noted Ben-David.

    That wouldn’t be a problem if GPUs had unlimited memory. But they don’t. Even the most advanced GPUs top out at around 288GB of high-bandwidth memory (HBM), and that space also has to hold the model itself.

    In real-world, multi-tenant inference environments, this becomes painful quickly. Workloads like code development or processing tax returns rely heavily on KV-cache for context.

    “If I'm loading three or four 100,000-token PDFs into a model, that's it — I've exhausted the KV cache capacity on HBM,” said Ben-David. This is what’s known as the memory wall. “Suddenly, what the inference environment is forced to do is drop data," he added.

    That means GPUs are constantly throwing away context they’ll soon need again, preventing agents from being stateful and maintaining conversations and context over time

    The hidden inference tax

    “We constantly see GPUs in inference environments recalculating things they already did,” Ben-David said. Systems prefill the KV cache, start decoding, then run out of space and evict earlier data. When that context is needed again, the whole process repeats — prefill, decode, prefill again. At scale, that’s an enormous amount of wasted work. It also means wasted energy, added latency, and degraded user experience — all while margins get squeezed.

    That GPU recalculation waste shows up directly on the balance sheet. Organizations can suffer nearly 40% overhead just from redundant prefill cycles This is creating ripple effects in the inference market.

    “If you look at the pricing of large model providers like Anthropic and OpenAI, they are actually teaching users to structure their prompts in ways that increase the likelihood of hitting the same GPU that has their KV cache stored,” said Ben-David. “If you hit that GPU, the system can skip the prefill phase and start decoding immediately, which lets them generate more tokens efficiently.”

    But this still doesn't solve the underlying infrastructure problem of extremely limited GPU memory capacity.

    Solving for stateful AI

    “How do you climb over that memory wall? How do you surpass it? That's the key for modern, cost- effective inferencing,” Ben-David said. “We see multiple companies trying to solve that in different ways.”

    Some organizations are deploying new linear models that try to create smaller KV caches. Others are focused on tackling cache efficiency.

    “To be more efficient, companies are using environments that calculate the KV cache on one GPU and then try to copy it from GPU memory or use a local environment for that,” Ben-David explained. “But how do you do that at scale in a cost-effective manner that doesn't strain your memory and doesn't strain your networking? That's something that WEKA is helping our customers with.”

    Simply throwing more GPUs at the problem doesn’t solve the AI memory barrier. “There are some problems that you cannot throw enough money at to solve," Ben-David said.

    Augmented memory and token warehousing, explained

    WEKA’s answer is what it calls augmented memory and token warehousing — a way to rethink where and how KV cache data lives. Instead of forcing everything to fit inside GPU memory, WEKA’s Augmented Memory Grid extends the KV cache into a fast, shared “warehouse” within its NeuralMesh architecture.

    In practice, this turns memory from a hard constraint into a scalable resource — without adding inference latency. WEKA says customers see KV cache hit rates jump to 96–99% for agentic workloads, along with efficiency gains of up to 4.2x more tokens produced per GPU.

    Ben-David put it simply: "Imagine that you have 100 GPUs producing a certain amount of tokens. Now imagine that those hundred GPUs are working as if they're 420 GPUs."

    For large inference providers, the result isn’t just better performance — it translates directly to real economic impact.

    “Just by adding that accelerated KV cache layer, we're looking at some use cases where the savings amount would be millions of dollars per day,” said Ben-David

    This efficiency multiplier also opens up new strategic options for businesses. Platform teams can design stateful agents without worrying about blowing up memory budgets. Service providers can offer pricing tiers based on persistent context, with cached inference delivered at dramatically lower cost.

    What comes next

    NVIDIA projects a 100x increase in inference demand as agentic AI becomes the dominant workload. That pressure is already trickling down from hyperscalers to everyday enterprise deployments— this isn’t just a “big tech” problem anymore.

    As enterprises move from proofs of concept into real production systems, memory persistence is becoming a core infrastructure concern. Organizations that treat it as an architectural priority rather than an afterthought will gain a clear advantage in both cost and performance.

    The memory wall is not something organizations can simply outspend to overcome. As agentic AI scales, it is one of the first AI infrastructure limits that forces a deeper rethink, and as Ben-David’s insights made clear, memory may also be where the next wave of competitive differentiation begins.

  • Z.ai's open source GLM-Image beats Google's Nano Banana Pro at complex text rendering, but not aesthetics Wed, 14 Jan 2026 20:59:00 GMT

    The two big stories of AI in 2026 so far have been the incredible rise in usage and praise for Anthropic's Claude Code and a similar huge boost in user adoption for Google's Gemini 3 AI model family released late last year — the latter of which includes Nano Banana Pro (also known as Gemini 3 Pro Image), a powerful, fast, and flexible image generation model that renders complex, text-heavy infographics quickly and accurately, making it an excellent fit for enterprise use (think: collateral, trainings, onboarding, stationary, etc).

    But of course, both of those are proprietary offerings. And yet, open source rivals have not been far behind.

    This week, we got a new open source alternative to Nano Banana Pro in the category of precise, text-heavy image generators: GLM-Image, a new 16-billion parameter open-source model from recently public Chinese startup Z.ai.

    By abandoning the industry-standard "pure diffusion" architecture that powers most leading image generator models in favor of a hybrid auto-regressive (AR) + diffusion design, GLM-Image has achieved what was previously thought to be the domain of closed, proprietary models: state-of-the-art performance in generating text-heavy, information-dense visuals like infographics, slides, and technical diagrams.

    It even beats Google's Nano Banana Pro on the shared by z.ai — though in practice, my own quick usage found it to be far less accurate at instruction following and text rendering (and other users seem to agree).

    But for enterprises seeking cost-effective and customizable, friendly-licensed alternatives to proprietary AI models, z.ai's GLM-Image may be "good enough" or then some to take over the job of a primary image generator, depending on their specific use cases, needs and requirements.

    The Benchmark: Toppling the Proprietary Giant

    The most compelling argument for GLM-Image is not its aesthetics, but its precision. In the CVTG-2k (Complex Visual Text Generation) benchmark, which evaluates a model's ability to render accurate text across multiple regions of an image, GLM-Image scored a Word Accuracy average of 0.9116.

    To put that number in perspective, Nano Banana 2.0 aka Pro—often cited as the benchmark for enterprise reliability—scored 0.7788. This isn't a marginal gain; it is a generational leap in semantic control.

    While Nano Banana Pro retains a slight edge in single-stream English long-text generation (0.9808 vs. GLM-Image's 0.9524), it falters significantly when the complexity increases.

    As the number of text regions grows, Nano Banana's accuracy remains in the 70s, whereas GLM-Image maintains >90% accuracy even with multiple distinct text elements.

    For enterprise use cases—where a marketing slide needs a title, three bullet points, and a caption simultaneously—this reliability is the difference between a production-ready asset and a hallucination.

    Unfortunately, my own usage of a demo inference of GLM-Image on Hugging Face proved to be less reliable than the benchmarks might suggest.

    My prompt to generate an "infographic labeling all the major constellations visible from the U.S. Northern Hemisphere right now on Jan 14 2026 and putting faded images of their namesakes behind the star connection line diagrams" did not result in what I asked for, instead fulfilling maybe 20% or less of the specified content.

    But Google's Nano Banana Pro handled it like a champ, as you'll see below:

    Of course, a large portion of this is no doubt due to the fact that Nano Banana Pro is integrated with Google search, so it can look up information on the web in response to my prompt, whereas GLM-Image is not, and therefore, likely requires far more specific instructions about the actual text and other content the image should contain.

    But still, once you're used to being able to type some simple instructions and get a fully researched and well populated image via the latter, it's hard to imagine deploying a sub-par alternative unless you have very specific requirements around cost, data residency and security — or the customizability needs of your organization are so great.

    Furthermore, Nano Banana Pro still edged out GLM-Image in terms of pure aesthetics — using the OneIG benchmark, Nano Banana 2.0 is at 0.578 vs. GLM-Image at 0.528 — and indeed, as the top header artwork of this article indicates, GLM-Image does not always render as crisp, finely detailed and pleasing an image as Google's generator.

    The Architectural Shift: Why "Hybrid" Matters

    Why does GLM-Image succeed where pure diffusion models fail? The answer lies in Z.ai’s decision to treat image generation as a reasoning problem first and a painting problem second.

    Standard latent diffusion models (like Stable Diffusion or Flux) attempt to handle global composition and fine-grained texture simultaneously.

    This often leads to "semantic drift," where the model forgets specific instructions (like "place the text in the top left") as it focuses on making the pixels look realistic.

    GLM-Image decouples these objectives into two specialized "brains" totaling 16 billion parameters:

    1. The Auto-Regressive Generator (The "Architect"): Initialized from Z.ai’s GLM-4-9B language model, this 9-billion parameter module processes the prompt logically. It doesn't generate pixels; instead, it outputs "visual tokens"—specifically semantic-VQ tokens. These tokens act as a compressed blueprint of the image, locking in the layout, text placement, and object relationships before a single pixel is drawn. This leverages the reasoning power of an LLM, allowing the model to "understand" complex instructions (e.g., "A four-panel tutorial") in a way diffusion noise predictors cannot.

    2. The Diffusion Decoder (The "Painter"): Once the layout is locked by the AR module, a 7-billion parameter Diffusion Transformer (DiT) decoder takes over. Based on the CogView4 architecture, this module fills in the high-frequency details—texture, lighting, and style.

    By separating the "what" (AR) from the "how" (Diffusion), GLM-Image solves the "dense knowledge" problem. The AR module ensures the text is spelled correctly and placed accurately, while the Diffusion module ensures the final result looks photorealistic.

    Training the Hybrid: A Multi-Stage Evolution

    The secret sauce of GLM-Image’s performance isn't just the architecture; it is a highly specific, multi-stage training curriculum that forces the model to learn structure before detail.

    The training process began by freezing the text word embedding layer of the original GLM-4 model while training a new "vision word embedding" layer and a specialized vision LM head.

    This allowed the model to project visual tokens into the same semantic space as text, effectively teaching the LLM to "speak" in images. Crucially, Z.ai implemented MRoPE (Multidimensional Rotary Positional Embedding) to handle the complex interleaving of text and images required for mixed-modal generation.

    The model was then subjected to a progressive resolution strategy:

    • Stage 1 (256px): The model trained on low-resolution, 256-token sequences using a simple raster scan order.

    • Stage 2 (512px - 1024px): As resolution increased to a mixed stage (512px to 1024px), the team observed a drop in controllability. To fix this, they abandoned simple scanning for a progressive generation strategy.

    In this advanced stage, the model first generates approximately 256 "layout tokens" from a down-sampled version of the target image.

    These tokens act as a structural anchor. By increasing the training weight on these preliminary tokens, the team forced the model to prioritize the global layout—where things are—before generating the high-resolution details. This is why GLM-Image excels at posters and diagrams: it "sketches" the layout first, ensuring the composition is mathematically sound before rendering the pixels.

    Licensing Analysis: A Permissive, If Slightly Ambiguous, Win for Enterprise

    For enterprise CTOs and legal teams, the licensing structure of GLM-Image is a significant competitive advantage over proprietary APIs, though it comes with a minor caveat regarding documentation.

    The Ambiguity: There is a slight discrepancy in the release materials. The model’s Hugging Face repository explicitly tags the weights with the MIT License.

    However, the accompanying GitHub repository and documentation reference the Apache License 2.0.

    Why This Is Still Good News: Despite the mismatch, both licenses are the "gold standard" for enterprise-friendly open source.

    • Commercial Viability: Both MIT and Apache 2.0 allow for unrestricted commercial use, modification, and distribution. Unlike the "open rail" licenses common in other image models (which often restrict specific use cases) or "research-only" licenses (like early LLaMA releases), GLM-Image is effectively "open for business" immediately.

    • The Apache Advantage (If Applicable): If the code falls under Apache 2.0, this is particularly beneficial for large organizations. Apache 2.0 includes an explicit patent grant clause, meaning that by contributing to or using the software, contributors grant a patent license to users. This reduces the risk of future patent litigation—a major concern for enterprises building products on top of open-source codebases.

    • No "Infection": Neither license is "copyleft" (like GPL). You can integrate GLM-Image into a proprietary workflow or product without being forced to open-source your own intellectual property.

    For developers, the recommendation is simple: Treat the weights as MIT (per the repository hosting them) and the inference code as Apache 2.0. Both paths clear the runway for internal hosting, fine-tuning on sensitive data, and building commercial products without a vendor lock-in contract.

    The "Why Now" for Enterprise Operations

    For the enterprise decision maker, GLM-Image arrives at a critical inflection point. Companies are moving beyond using generative AI for abstract blog headers and into functional territory: multilingual localization of ads, automated UI mockup generation, and dynamic educational materials.

    In these workflows, a 5% error rate in text rendering is a blocker. If a model generates a beautiful slide but misspells the product name, the asset is useless. The benchmarks suggest GLM-Image is the first open-source model to cross the threshold of reliability for these complex tasks.

    Furthermore, the permissive licensing fundamentally changes the economics of deployment. While Nano Banana Pro locks enterprises into a per-call API cost structure or restrictive cloud contracts, GLM-Image can be self-hosted, fine-tuned on proprietary brand assets, and integrated into secure, air-gapped pipelines without data leakage concerns.

    The Catch: Heavy Compute Requirements

    The trade-off for this reasoning capability is compute intensity. The dual-model architecture is heavy. Generating a single 2048x2048 image requires approximately 252 seconds on an H100 GPU. This is significantly slower than highly optimized, smaller diffusion models.

    However, for high-value assets—where the alternative is a human designer spending hours in Photoshop—this latency is acceptable.

    Z.ai also offers a managed API at $0.015 per image, providing a bridge for teams who want to test the capabilities without investing in H100 clusters immediately.

    GLM-Image is a signal that the open-source community is no longer just fast-following proprietary labs; in specific, high-value verticals like knowledge-dense generation, they are now setting the pace. For the enterprise, the message is clear: if your operational bottleneck is the reliability of complex visual content, the solution is no longer necessarily a closed Google product—it might be an open-source model you can run yourself.

  • AI agents can talk — orchestration is what makes them work together Wed, 14 Jan 2026 19:00:00 GMT

    Rather than asking how AI agents can work for them, a key question in enterprise is now: Are agents playing well together? 

    This makes orchestration across multi-agent systems and platforms a critical concern — and a key differentiator. 

    “Agent-to-agent communications is emerging as a really big deal,” G2’s chief innovation officer Tim Sanders told VentureBeat. “Because if you don't orchestrate it, you get misunderstandings, like people speaking foreign languages to each other. Those misunderstandings reduce the quality of actions and raise the specter of hallucinations, which could be security incidents or data leakage.”

    Allowing agents to talk and coordinate

    Orchestration to this point has largely been around data, but that’s quickly turning to action. “Conductor-like solutions” are increasingly bringing together agents, robotic process automation (RPA), and data repositories. Sanders likened the progression to that of answer engine optimization, which initially began with monitoring and now creates bespoke content and code. 

    “Orchestration platforms coordinate a variety of different agentic solutions to increase the consistency of outcomes,” he said. 

    Early providers include Salesforce MuleSoft, UiPath Maestro, and IBM Watsonx Orchestrate. These “phase one” software-based observability dashboards help IT leaders see all agentic actions across an enterprise. 

    The critical element of risk management

    But coordination can only add so much value; these platforms will morph into technical risk management tools that provide greater quality control. This could include, for instance, agent assessments, policy recommendation and proactive scoring (such as, how reliable agents are when they call on enterprise tools, or how often they hallucinate and when). 

    Enterprise leaders have become wary of relying on vendors to minimize risks and errors; many IT decision-makers, in fact, do not trust a vendor's statements about the reliability of their agents, he said. 

    Third-party tools are beginning to bridge the gap and automate tedious guardrail processes and escalation tickets. Teams are already experiencing “ticket exhaustion” in semi-automated systems, where agents hit guardrails and require human permission to proceed.

    As an example: The loan process at a bank requires 17 steps for approval, and an agent keeps interrupting human workflows with approval requests when it runs into established guardrails.

    Third-party orchestration platforms can manage these tickets and nay, yay, or even challenge the need for approval altogether. They can eventually eliminate the need for persistent human-in-the-loop oversight so organizations can experience “true velocity gains” measured not in percentages but in multiples (that is, 3X versus 30%).

    “Where it goes from there is remote management of the entire agentic process for organizations,” Sanders said. 

    ‘Human-on-the-loop’ versus ‘human-in-the-loop’ 

    In another critical evolution in the agentic era, human evaluators will become designers, moving from human-in-the-loop to human-on-the-loop, according to Sanders. That is: They will begin designing agents to automate workflows. 

    Agent builder platforms continue to innovate their no-code solutions, Sanders said, meaning nearly anyone can now stand up an agent using natural language. “This will democratize agentic AI, and the super skill will be the ability to express a goal, provide context and envision pitfalls, very similar to a good people manager today.”

    What enterprise leaders should be doing now

    Agent-first automation stacks “dramatically outperform” hybrid automation stacks in almost every attribute, he noted: satisfaction, quality of actions, security, cost savings.

    Organizations should begin “expeditious programs” to infuse agents across workflows, especially with highly repetitive work that poses bottlenecks. Likely at first, there will be a strong human-in-the-loop element to ensure quality and promote change management. 

    “Serving as an evaluator will strengthen the understanding of how these systems work,” Sanders said, “and eventually enable all of us to operate upstream in agentic workflows instead of downstream.” 

    IT leaders should take inventory today of all the different elements of their automation stack. Whether these elements are rules-based automation, RPA, or agentic automation, they must learn everything going on in the organization to optimally use emerging orchestration platforms.

    “If they don't, there could actually be dis-synergies across organizations where old school technology and cutting edge technology clash at the point of delivery, oftentimes customer-facing,” Sanders said. “You can't orchestrate what you can't see clearly.”



Techradar



TechNode

  • Xiaomi, Oppo, Vivo, Transsion cut 2026 smartphone shipment forecasts Fri, 16 Jan 2026 10:00:32 +0000
    China’s smartphone makers Xiaomi, Oppo, Vivo and Transsion have lowered their full-year shipment forecasts amid rising memory prices in the supply chain, according to Jiemian News. Xiaomi and Oppo have cut their projections by more than 20%, Vivo by nearly 15%, and Transsion to below 70 million units. The reductions mainly affect mid- to low-end […]
  • From lab to living room: how CSOT turns research into reliable displays Fri, 16 Jan 2026 08:43:17 +0000
    As consumer electronics and smart devices become more common, display technology plays an increasingly important role in electronic devices. Chinese company TCL China Star Optoelectronics Technology (CSOT) has conducted research in LCD, OLED, and emerging display technologies, making notable advancements across these areas. Today, we’re going to take a closer look at CSOT’s innovations in […]
  • Apple expands Apple Pay cross-border payments for mainland China users Fri, 16 Jan 2026 07:13:38 +0000
    On Thursday, Apple announced an expansion of Apple Pay’s cross-border payment services for users in mainland China. The service is now available to users with Visa credit cards issued by eight Chinese banks, including Industrial and Commercial Bank of China, Bank of China and Agricultural Bank of China, as well as Visa debit cards issued […]
  • Alibaba’s Qwen app links Taobao, Alipay in end-to-end AI shopping push Fri, 16 Jan 2026 02:11:12 +0000
    Alibaba Group said on Thursday that its Qwen app has been fully integrated with key services across its ecosystem, including Taobao, Alipay and travel platform Fliggy, enabling users to complete tasks such as ordering food, booking flights and shopping through a single AI-driven interface. The company said users can carry out the entire process — […]
  • ByteDance’s next-gen AI earphones to be made by Goertek Thu, 15 Jan 2026 03:08:07 +0000
    ByteDance is developing a next-generation version of its Doubao AI-powered earphones, with Chinese manufacturer Goertek set to handle contract manufacturing, according to supply chain sources cited by Blue Whale Tech. The sources said Goertek has established a dedicated business group to oversee its cooperation with ByteDance, reflecting the strategic importance the supplier places on the […]
  • AI-led demand signals longer semiconductor upcycle into 2026 and beyond Thu, 15 Jan 2026 02:51:31 +0000
    At the UBS Greater China Conference 2026 held in Shanghai on Wednesday, Jimmy Yu, China Technology Semiconductor Analyst at UBS Securities, shared his latest views on global and China’s semiconductor industry cycles, AI-driven growth dynamics, and investment themes. He said that amid the rapid development of AI, the global semiconductor industry is entering a new […]
  • Ant International, Google to collaborate on AI commerce protocol Tue, 13 Jan 2026 06:34:33 +0000
    Ant International and Google are working together on a new open standard aimed at enabling artificial intelligence-driven commerce, according to a reporter from China Star Market. The two companies are jointly promoting the Universal Commerce Protocol (UCP), an emerging framework designed to support so-called Agentic Commerce, in which AI agents can autonomously assist users throughout […]
  • Chinese Neuralink Rival BrainCo Files for Hong Kong IPO, Sources Say Tue, 13 Jan 2026 02:52:57 +0000
    BrainCo has confidentially filed for an initial public offering in Hong Kong, Bloomberg reported. The company is part of the “Hangzhou Six,” a group of emerging tech startups from the city that includes DeepSeek. Sources said the listing could raise several hundred million dollars, with CICC and UBS Group working on the share sale. Founded […]
  • Zhipu AI partners with Didi to develop AGI-based mobility agents Tue, 13 Jan 2026 02:32:09 +0000
    Chinese AI firm Zhipu AI and ride-hailing giant Didi have agreed to form a strategic partnership focused on AGI (Artificial General Intelligence) and agent-based applications for mobility, the companies said. The collaboration will centre on three areas, including testing the deployment of intelligent agents in complex mobility scenarios to improve user intent alignment and reasoning […]
  • Honor reportedly partnering with Pop Mart, designer toy aesthetics coming to smartphones Mon, 12 Jan 2026 07:45:39 +0000
    Honor is preparing to launch an IP co-branded collaboration with trendy toy maker Pop Mart, Chinese media outlet China Star Market reported. Pop Mart has previously said it has no plans to develop smartphones, with the partnership instead focusing on cross-industry IP integration. Pop Mart has posted strong sales from earlier IP tie-ups, including co-branded […]
  • X Square Robot secures $140 million in funding from ByteDance, Sequoia, and others Mon, 12 Jan 2026 05:56:06 +0000
    Embodied intelligence company X Square Robot announced today that it has completed one billion yuan ($140 million) A++ funding round. The round was led by top-tier investors including ByteDance, Sequoia China, and Shenzhen Capital Group, along with local government-backed platforms. Notably, this marks the first investment by Shenzhen Capital’s newly established AI fund. Founded in […]
  • Dead or Alive app designed for people living alone sparks debate in China Mon, 12 Jan 2026 02:41:22 +0000
    A mobile app called Dead or Alive has gone viral on Chinese social media platform Weibo, drawing widespread attention for its focus on safety for people living alone. Developed by three post-1995-born programmers in less than a month at a cost of just over 1,000 yuan ($140), the app is designed to monitor users’ well-being. […]
  • MiHoYo-backed AI firm MiniMax jumps on Hong Kong debut, market value tops $11.5 billion Fri, 09 Jan 2026 09:51:34 +0000
    Shares of MiniMax, an AI firm backed by Chinese game developer MiHoYo, surged more than 70% in their Hong Kong trading debut today, briefly pushing the company’s market capitalisation above HK$90 billion ($11.5 billion). MiniMax, founded by former SenseTime vice president Yan Junjie, listed on the main board of the Hong Kong Stock Exchange under […]
  • China’s AgiBot leads global humanoid robot shipments in 2025, Omdia says Fri, 09 Jan 2026 06:19:50 +0000
    Chinese humanoid robot maker AgiBot ranked first globally by shipment volume in 2025, delivering 5,168 units and accounting for 39% of the global market, according to a report released by market research firm Omdia. In its report, General-Purpose Embodied Robotics Market Radar, Omdia said the global humanoid robot market entered a phase of rapid growth […]
  • MiniMax lists in Hong Kong as largest IPO among AI foundation model companies Fri, 09 Jan 2026 05:39:51 +0000
    Shanghai-based AI company MiniMax listed on the Hong Kong Stock Exchange on Jan. 9. Shares surged more than 50% during intraday trading, pushing the company’s market capitalization above HK$76.3 billion (about $9.8 billion). The company raised about HK$5.54 billion after pricing its global offering at HK$165 per share, the top of the indicated range. The […]
  • TSMC slows launch of new 3nm projects, encourages clients to consider 2nm Fri, 09 Jan 2026 03:33:11 +0000
    TSMC has temporarily halted the launch of new 3nm chip projects and raised prices for the advanced node, as demand from AI and high-end computing customers continues to strain capacity, chipmakers familiar with the matter said. Existing 3nm capacity has been fully taken up by orders for AI GPUs, cloud data centre application-specific integrated circuits […]
  • BYD Files 4 Models Under New “Linghui” Brand, Speculated for Ride-Hailing Fri, 09 Jan 2026 03:32:54 +0000
    Chinese EV maker BYD has appeared in the 403rd product catalog released by China’s Ministry of Industry and Information Technology (MIIT) with a new sub-brand, “Linghui” (领汇, our translation). The filing showcases a distinct new logo for the marque, which market observers speculate may specifically target the ride-hailing sector to differentiate fleet sales from consumer […]
  • Zhipu AI Becomes the World’s First Public Company Centered on AGI Foundation Models with Hong Kong Listing Thu, 08 Jan 2026 08:02:57 +0000
    Chinese AI company Zhipu AI listed on the Hong Kong Stock Exchange today, opening at HK$120 per share with a market capitalization exceeding HK$52.8 billion (approx. $6.8 billion). The listing makes Zhipu the world’s first publicly listed company with AGI foundation models as its core business. In an internal letter released shortly after the listing, […]
  • ByteDance denies entering car manufacturing despite automotive tech partnerships Thu, 08 Jan 2026 06:59:51 +0000
    ByteDance on Wednesday denied market speculation that it plans to produce vehicles in partnership with automakers. “The company has no plans to manufacture vehicles,” a ByteDance spokesperson told multiple media outlets. This follows a similar clarification in July 2025, when the company dismissed reports that its Volcano Engine division was preparing the Doubao Car. While […]
  • CES 2026 Live: Vbot Steals the Show, ‘Vbot SuperDog’ Poised to Become One of the Year’s Hottest Tech Gadgets Thu, 08 Jan 2026 05:15:48 +0000
    On January 6, the highly anticipated CES 2026 kicked off in Las Vegas, marking the largest and most AI-integrated exhibition in the event’s history. Amidst the technological frenzy, one booth stood out distinctively: Vbot (Vbot维他动力). Unlike traditional exhibits where hardware is fenced off for display, Vbot’s robots walked freely among the crowds, marking a shift […]
  • Goose Goose Duck mobile crashes on day one of China release, prompting repeated apologies Thu, 08 Jan 2026 03:57:08 +0000
    The China mobile version of Goose Goose Duck launched across all platforms on Wednesday, but was almost immediately hit by server crashes as an overwhelming number of players flooded in on day one. Many users reported disconnections and difficulties logging in, while scalpers were seen reselling game IDs at inflated prices. The issue quickly set […]
  • CES 2026: small-size autonomous vehicle tackles the last 100 meters Wed, 07 Jan 2026 09:31:23 +0000
    Neolix, the Chinese autonomous vehicle maker, showcased its next-generation AI logistics solutions at CES 2026, featuring the X1 — a small, agile vehicle designed for last-mile deliveries. Capable of navigating sidewalks, accessing office lobbies, and taking elevators, the X1 can reach customers’ doors even in tight urban spaces. Part of Neolix’s full RoboVan lineup, which […]
  • Xiaomi Opens Pre-Orders for Updated SU7: Standard LiDAR & Range Up to 902km Wed, 07 Jan 2026 08:49:26 +0000
    Xiaomi began pre-orders today for the 2026 SU7, listing three specific trims: Standard (229,900 RMB / ~$31,700), Pro (259,900 RMB / ~$35,800), and Max (309,900 RMB / ~$42,700). Performance hardware sees a major baseline upgrade. All models now feature LiDAR, the V6s Plus motor, four-piston front calipers, and a staggered tire setup (245mm front / […]
  • CES 2026: This Concept Robot Vacuum Grows Legs to Climb Stairs Wed, 07 Jan 2026 08:27:49 +0000
    You might finally stop carrying your vacuum between floors. Dreame showcased its Cyber X concept, a cleaner equipped with retractable rubber tracks that allow it to autonomously navigate stairs and multi-level terrain. Originally debuted at IFA, the device can tackle steps up to 25cm high and inclines of 42 degrees. It functions as a standard […]
  • CES 2026: first phone-free AI glasses with eSIM and 4G connectivity unveiled Wed, 07 Jan 2026 08:14:49 +0000
    At CES 2026, Chinese tech company RayNeo introduced the X3 Pro Project eSIM AR glasses, which it says are the first AR glasses for consumers that can work independently. With a built-in eSIM and 4G connection, the glasses allow you to make calls, chat with AI in multiple ways, translate languages instantly, and stream music […]
  • CES 2026: This Motorized Dock Turns Your iPhone Into a Tracking AI Robot Wed, 07 Jan 2026 06:31:28 +0000
    You can now transform your iPhone into an interactive desktop companion that physically tracks your movement, rather than just charging it. Unveiled at CES 2026, the DeskMate is a motorized desktop dock equipped with 3 USB-C ports and 1 USB-A port. When attached via MagSafe, it uses your phone’s screen and camera to display Pixar-style […]
  • CES 2026: world’s first wheel-legged robot vacuum designed to tackle stairs and multi-level homes Wed, 07 Jan 2026 06:06:47 +0000
    Roborock unveiled a concept device dubbed G-Rover at CES 2026, pitching it as the world’s first wheel-legged robot vacuum designed to tackle stairs and multi-level homes. The prototype features a dual wheel-leg architecture, with each leg able to lift independently and adjust its height. Paired with AI algorithms and 3D spatial perception, the system enables […]
  • CES 2026: AI-powered 3D digital avatars to support refereeing at FIFA World Cup 2026 Wed, 07 Jan 2026 03:13:33 +0000
    Lenovo and FIFA announced a series of joint technology solutions at Lenovo’s Tech World event during CES 2026, aiming to deliver an AI-powered FIFA World Cup at the 2026 tournament. Under the partnership, digital avatars will be integrated into refereeing technologies and match broadcasts at the FIFA World Cup 2026. Lenovo and FIFA are using […]
  • CES 2026: smart refrigerator with instant sparkling water system unveiled Tue, 06 Jan 2026 08:03:12 +0000
    Chinese tech firm Dreame Technology showcased a smart refrigerator equipped with an instant sparkling water system at the CES 2026 technology show. The refrigerator features a built-in carbonation module that dispenses sparkling water through a dedicated outlet, offering multiple carbonation levels without the need for external devices, the company said. It also includes an ice-making […]
  • CES 2026: limited-edition gaming tablet blends AI computing with Kojima Productions design Tue, 06 Jan 2026 05:49:56 +0000
    Asus’ gaming brand Republic of Gamers (ROG) said at CES 2026 it is collaborating with Japanese game studio Kojima Productions to launch a limited-edition version of its Flow Z13 2-in-1 gaming tablet, marking a rare crossover between gaming hardware and auteur-led game design. The ROG Flow Z13-KJP is built around Kojima Productions’ Ludens concept, with […]



How Technology Works demystifies the machinery that keeps the modern world going, from simple objects such as zip fasteners and can openers to the latest, most sophisticated devices of the information age, including smartwatches, personal digital assistants, and driverless cars. #ad