As India pitches itself as a global hub for applied artificial intelligence, OpenAI has partnered with Pine Labs to integrate AI-driven reasoning into the fintech firm’s payments stack, automating settlement and invoicing workflows in a move the companies say could help accelerate AI-led commerce in India.
The partnership will see Pine Labs embed OpenAI’s application programming interfaces — software tools that let companies plug AI into their existing systems — within its payments and commerce infrastructure, the companies said on Thursday, all with the aim of enabling AI-assisted settlement, reconciliation, and invoicing workflows.
The deal underscores OpenAI’s broader push to expand its footprint in India, one of its fastest-growing markets, as it looks to move beyond being known primarily as the maker of ChatGPT and embed its technology into education, enterprise, and infrastructure. Earlier this week, OpenAI partnered with leading Indian engineering, medical, and design institutions to bring AI tools into higher education, betting that India’s large developer base and more than a billion internet users will play a central role in the next phase of AI adoption.
Pine Labs is already using AI internally to automate parts of its settlement and reconciliation process, cutting the time it takes to clear daily settlements from hours to minutes, according to Chief executive B Amrish Rau. The Noida-based company previously relied on manual checks by dozens of employees to process funds from multiple banks before markets opened each day, a workflow that is now largely handled by AI-driven systems, he said in an interview.
Advertisement
For Pine Labs, the partnership is intended to extend those AI-driven efficiencies beyond internal operations to merchants and corporate clients, starting with business-to-business use cases such as invoice processing, settlements and payments orchestration, Rau told TechCrunch. He noted the company sees faster adoption in B2B workflows, where AI agents can handle large volumes of repetitive financial tasks under predefined rules, before similar capabilities reach consumer-facing payments.
“People talk about retail AI, but the bigger impact of all of this is really efficiency improvement, especially in B2B,” Rau said. “If you look at invoicing and settlement, those are workflows where agents can actually drive the process end to end, and that’s where adoption can happen faster.”
The rollout of more autonomous, agent-led payment workflows will move faster in overseas markets where regulations already allow such transactions, Rau said, while India is likely to see a more gradual adoption focused on AI-assisted commerce rather than fully agent-initiated payments. He said that Pine Labs is already prototyping agent-driven payments in parts of the Middle East and Southeast Asia, even as Indian regulations require tighter controls on how payments are authorized.
Techcrunch event
Advertisement
Boston, MA | June 23, 2026
For OpenAI, the partnership offers a route deeper into India’s payments and enterprise ecosystem as it looks to move beyond consumer-facing tools and embed its models into high-volume, regulated workflows. Rau said the collaboration is aimed at increasing merchant stickiness and expanding Pine Labs’ role from a payments processor to a broader commerce platform, with higher transaction volumes over time translating into incremental revenue.
Advertisement
Pine Labs says it works with more than 980,000 merchants, 716 consumer brands, and 177 financial institutions, and has processed over 6 billion cumulative transactions valued at over ₹11.4 trillion (about $126 billion), per its prospectus published last year. The fintech operates across 20 countries, including Malaysia, Singapore, Australia, parts of Africa, the UAE, and the U.S., giving the OpenAI partnership reach across both Indian and international markets.
Rau said the partnership does not involve revenue sharing between the two companies, with Pine Labs not taking a cut if its merchants choose to embed OpenAI’s tools. “We’ve kept it completely independent of each other — anything related to payment and payment services, we will get the benefit of it, and anything related to OpenAI revenues will go to them,” he said.
The arrangement, Rau added, is also non-exclusive. He compared it to OpenAI’s partnership with Stripe in the U.S. and said Pine Labs remains open to working with other AI providers.
Rau said Pine Labs is building additional security and compliance layers around AI-driven workflows to ensure that sensitive merchant and consumer transaction data remains protected, as the company integrates AI more deeply into its payments systems. He said the focus is on ensuring transactions remain secure and compliant even as more workflows are automated by AI.
The new announcement comes as India hosts its AI Impact Summit in New Delhi, where global AI companies including OpenAI, Anthropic, and Google are showcasing their latest capabilities alongside Indian startups demonstrating AI applications aimed at large-scale deployment across sectors such as finance, healthcare, and education.
On the podcast, [Tom] and I were talking about the new generation of smartphones which are, at least in terms of RAM and CPU speed, on par with a decent laptop computer. If so, why not just add on a screen, keyboard, and mouse and use it as your daily driver? That was the question posed by [ETA Prime] in a video essay and attempt to do so.
Our consensus was that it’s the Android operating system holding it back. Some of the applications you might want to run just aren’t there, and on the open side of the world, even more are missing. Is the platform usable if you can’t get the software you need to get your work done?
Looking back, these have been the same stumbling blocks for the last decade. In 2018, I was wondering aloud why we as a community don’t hack on cell phones, and the answer then was the same as it is now – the software is not friendly to our kind. You can write phone apps, and I have tried to do so, but it’s just not fun.
Advertisement
The polar opposites of the smartphone-as-computer are no strangers in our community. I’m thinking of the Linux single-board computers, or even something like a Steam Deck, all of which are significantly less powerful spec-wise than a flagship cell phone, but which are in many ways much more suitable for hacking. Why? Because they make it easy to do the things that we like to do. They’re designed to be fun computers, and so we use them.
So for me, a smartphone isn’t a computer, but oddly enough it’s not because of the hardware. It’s because what I want out of a computer is more than Turing completeness. What I want is the fun and the freedom of computering.
A new NYT Connections puzzle appears at midnight each day for your time zone – which means that some people are always playing ‘today’s game’ while others are playing ‘yesterday’s’. If you’re looking for Saturday’s puzzle instead then click here: NYT Connections hints and answers for Saturday, March 14 (game #1007).
Good morning! Let’s play Connections, the NYT’s clever word game that challenges you to group answers in various categories. It can be tough, so read on if you need Connections hints.
What should you do once you’ve finished? Why, play some more word games of course. I’ve also got daily Strands hints and answers and Quordle hints and answers articles if you need help for those too, while Marc’s Wordle today page covers the original viral word game.
Advertisement
SPOILER WARNING: Information about NYT Connections today is below, so don’t read on if you don’t want to know the answers.
Article continues below
NYT Connections today (game #1008) – today’s words
(Image credit: New York Times)
Today’s NYT Connections words are…
GOSSIP
HOUND
NEW
TOY
SPORTING
WORKING
BOOK
FAIR
SHADOW
GONE
BIKE
TRACK
SQUARE
VIDEO GAME
TAIL
HONEST
NYT Connections today (game #1008) – hint #1 – group hints
What are some clues for today’s NYT Connections groups?
YELLOW: Following
GREEN: Honorable behavior
BLUE: What to buy a 10-year-old
PURPLE: Screen arts featuring “a young woman”
Need more clues?
We’re firmly in spoiler territory now, but read on if you want to know what the four theme answers are for today’s NYT Connections puzzles…
Advertisement
Sign up for breaking news, reviews, opinion, top tech deals, and more.
NYT Connections today (game #1008) – hint #2 – group answers
What are the answers for today’s NYT Connections groups?
YELLOW: PURSUE
GREEN: SPORTSMANLIKE
BLUE: CLASSIC KID GIFTS
PURPLE: “____ GIRL” TITLES
Right, the answers are below, so DO NOT SCROLL ANY FURTHER IF YOU DON’T WANT TO SEE THEM.
Advertisement
NYT Connections today (game #1008) – the answers
(Image credit: New York Times)
The answers to today’s Connections, game #1008, are…
BLUE: CLASSIC KID GIFTS BIKE, BOOK, TOY, VIDEO GAME
PURPLE: “____ GIRL” TITLES GONE, GOSSIP, NEW, WORKING
My rating: Easy
My score: Perfect
Seeing TOY and BOOK alongside FAIR temporarily sidelined me into attempting a fair group (are VIDEO GAME fairs a thing?) and GOSSIP and HOUND seemed to hint at another collection, but fortunately I quickly moved on from these ideas.
In most groups there is always one tile that you feel less certain about than the others and in PURSUE it was HOUND, which sounded closer to up-close harassment than the covert activities suggested by SHADOW, TAIL, and TRACK.
Advertisement
For SPORTSMANLIKE the tricky fourth tile was SQUARE (isn’t this what you’d call an uncool sportsman?), while for CLASSIC KID GIFTS it was most definitely BOOK.
Yesterday’s NYT Connections answers (Saturday, March 14, game #1007)
YELLOW: HYPNOTIC STATE DREAM, HAZE, SPELL, TRANCE
GREEN: STARTING WITH PREFIXES MEANING “TWO” BINARY, DIOXIDE, DUOLINGO, TWILIGHT
PURPLE: ENDING IN FEMALE ANIMALS HOOTENANNY, LICHEN, MOSCOW, NIGHTMARE
What is NYT Connections?
NYT Connections is one of several increasingly popular word games made by the New York Times. It challenges you to find groups of four items that share something in common, and each group has a different difficulty level: green is easy, yellow a little harder, blue often quite tough and purple usually very difficult.
On the plus side, you don’t technically need to solve the final one, as you’ll be able to answer that one by a process of elimination. What’s more, you can make up to four mistakes, which gives you a little bit of breathing room.
Advertisement
It’s a little more involved than something like Wordle, however, and there are plenty of opportunities for the game to trip you up with tricks. For instance, watch out for homophones and other word games that could disguise the answers.
It’s playable for free via the NYT Games site on desktop or mobile.
We have all heard about model context protocol (MCP) in the context of artificial intelligence. In this article, we will dive into what MCP is and why it is becoming more important by the day. When APIs are already available then why do we need MCP? Although we have seen a large rise in popularity of MCP, is there staying power in this new protocol? In the first section, we will look at the parallels between APIs and MCP and then start to explore what sets it apart.
From APIs to model context protocol
A single isolated computer is limited in the amount of data that it can access and that has a direct impact on its usability. APIs were created to enable data transfer between systems. Just like APIs, Model Context Protocol (MCP) is the protocol for communication between AI agents that are using large language models (LLMs). APIs are primarily written for developers while MCP servers are created for AI agents (Johnson, 2025).
What is MCP?
MCP was introduced by Anthropic on November 25, 2024 as an open source standard to enable communication between AI assistants and external data sources. AI agents are constrained by the fragmentation of data in isolated systems (Anthropic, 2024). The protocol defines how agents can interact with external systems, elicit user input and enable automated agents.
At its core MCP utilizes the client server model and there are three main features for clients and servers.
To keep this article concise, focus will be kept on the most important feature of both client and server. For MCP servers, tools are the primary way to perform complex tasks and clients utilize elicitation to enable a two way communication between the agent and the user.
Instead of explicitly calling APIs, agents select and use the appropriate tools (functions) based on the input they receive from the user. If a tool requires certain parameters the agent will use elicitation to get the data from the user. This allows for a more responsive workflow where two way communication between LLM and the user is possible.
Advertisement
Why do we need MCP now?
A very valid question to ask is if APIs are already present then why is there a need for MCP? APIs are designed to connect fragmented data systems and SaaS applications already enable a two way communication with a user. So, why do we need MCP now?
The main need for MCP is that the user of external data has changed from developers to AI agents. A developer will usually program an application using APIs that behaves in a deterministic fashion. Whereas, AI agents will use the user prompt and make autonomous decisions to execute on the user request. By nature, the execution of a workflow by an AI agent is not deterministic.
APIs are a machine-executable contract which acts in a deterministic fashion. APIs work if the users of APIs know what action needs to be taken next (Posta, 2025). AI agents run on top of probabilistic LLMs which do not consistently deliver repeatable results across all tasks (Atil, 2024). Variance in a LLM’s response is expected and this poses a problem for autonomous execution.
MCP to the rescue
MCP solves the problem of variance in agent execution by providing high level abstraction that wraps functionality rather than API endpoints. Tools enable LLM models to perform actions like searching for a flight, booking a calendar and more (Understanding MCP Servers, 2026).
Advertisement
One common misconception for tools is that they are just an abstraction over existing API calls. Tools are not designed to be an abstraction over API calls but rather abstraction over functionality. If a lot of APIs are just exposed as tools it will increase the cost and context size for the agent which is not ideal (Johnson, 2025).
A tool may include multiple API calls in its implementation to achieve the desired outcome. An agent will review the list of available tools to automatically select the most appropriate tools and determine the appropriate order of execution.
MCP adoption boom
Since its release in 2024 MCP has seen a steady rise in popularity. The following chart from Google Trends showcases the relative interest in MCP since its launch.
A lot of companies have launched their own MCP servers to facilitate building autonomous agents. As of February 2026, the official MCP registry has over 6400 MCP servers already registered. This number of MCP servers is only expected to grow in the near future. The official registry for MCP servers is still in preview and the ecosystem has grown massively in less than a year.
Advertisement
Other major players in the market have adopted MCP and added support to their clients. OpenAI added MCP support to ChatGPT in March and Google added support a few weeks later in April 2025. This showcases the staying power of the protocol and the fast pace of adoption.
What lies ahead?
MCP is still in the early stages of widespread adoption where a lot of applications need to mature and start hitting production. Leonardo Pineryo from Pento AI summarized it the best “MCP’s first year transformed how AI systems connect to the world. Its second year will transform what they can accomplish” (2025).
Guardrails around tools is an area that will see further development as trust is one of the biggest concerns with AI agents. With better guardrails in the tools, an AI agent can be allowed to perform with more autonomy. Over the next year, MCP is certain to see continued growth, both in the sophistication of its capabilities and the volume of its application.
A new NYT Strands puzzle appears at midnight each day for your time zone – which means that some people are always playing ‘today’s game’ while others are playing ‘yesterday’s’. If you’re looking for Saturday’s puzzle instead then click here: NYT Strands hints and answers for Saturday, March 14 (game #741).
Strands is the NYT’s latest word game after the likes of Wordle, Spelling Bee and Connections – and it’s great fun. It can be difficult, though, so read on for my Strands hints.
Want more word-based fun? Then check out my NYT Connections today and Quordle today pages for hints and answers for those games, and Marc’s Wordle today page for the original viral word game.
Advertisement
SPOILER WARNING: Information about NYT Strands today is below, so don’t read on if you don’t want to know the answers.
Article continues below
NYT Strands today (game #742) – hint #1 – today’s theme
What is the theme of today’s NYT Strands?
• Today’s NYT Strands theme is… Best of all
NYT Strands today (game #742) – hint #2 – clue words
Play any of these words to unlock the in-game hints system.
Advertisement
GEAR
STAGE
SNOW
SNAG
CARER
TURN
NYT Strands today (game #742) – hint #3 – spangram letters
How many letters are in today’s spangram?
• Spangram has 12 letters
NYT Strands today (game #742) – hint #4 – spangram position
What are two sides of the board that today’s spangram touches?
First side: left, 5th row
Last side: right, 5th row
Advertisement
Right, the answers are below, so DO NOT SCROLL ANY FURTHER IF YOU DON’T WANT TO SEE THEM.
Advertisement
NYT Strands today (game #742) – the answers
(Image credit: New York Times)
The answers to today’s Strands, game #742, are…
DIRECTOR
SOUND
ACTOR
ACTRESS
PICTURE
SONG
SPANGRAM: ACADEMYAWARD
My rating: Easy
My score: Perfect
This Sunday the world will be tuned into the 98th ACADEMY AWARD ceremony in Los Angeles, as the 3,141st to the 3,165th statuettes are handed out to the “best of all”.
Sign up for breaking news, reviews, opinion, top tech deals, and more.
Despite the many things I don’t like about it (the industry thank you list speeches, mainly), I always really enjoy watching the Oscars — although in the UK it does require staying awake until 3.30am — and this year I’ll be hoping The Secret Agent gets some recognition.
I digress, with little in the way of mystery this edition of Strands was fairly straightforward. My only error was adding an S to the end of the Spangram and having to connect it all over again after it didn’t turn yellow.
Advertisement
Yesterday’s NYT Strands answers (Saturday, March 14, game #741)
VENT
CRUST
FILLING
LATTICE
GLAZE
FRUIT
EDGES
SPANGRAM: HAPPYPIDAY
What is NYT Strands?
Strands is the NYT’s not-so-new-any-more word game, following Wordle and Connections. It’s now a fully fledged member of the NYT’s games stable that has been running for a year and which can be played on the NYT Games site on desktop or mobile.
I’ve got a full guide to how to play NYT Strands, complete with tips for solving it, so check that out if you’re struggling to beat it each day.
AI services may not stay cheap for long, as companies like OpenAI and Anthropic are currently subsidizing usage to rapidly grow market share. As these companies move toward profitability and potential IPOs, Axios reports that investors will likely push them to increase prices and improve margins. An anonymous reader shares an excerpt from the report: Flashback: Silicon Valley has seen this movie before. The so-called “millennial lifestyle subsidy” meant VC money helped underwrite cheap Uber rides and DoorDash deliveries. Before that, Amazon built its base with low prices, free shipping and, for years, no sales tax in most states. Eventually, all of these companies had to charge enough to cover costs — and make a profit.
Follow the money: The current iteration of AI subsidies won’t last forever. Both OpenAI and Anthropic are widely expected to go public. Public investors will demand earnings growth and expanding margins. Even as chips get more efficient, total spending keeps rising. Labs need more capacity, more upgrades and more supply to meet demand.
The bottom line: The costs of AI will keep going down. But total spend from customers will need to keep going up if AI companies are going to become profitable and investors are ever going to get returns on their massive investments.
The feature, announced at SXSW by co-CEO Gustav Söderström, lets Premium listeners see and shape the data model powering their recommendations, starting with a beta rollout in New Zealand
For a decade, Spotify’s recommendation engine has worked largely in silence. It watched what you played, noted what you skipped, inferred meaning from the time of day and the tempo of your commute, and it never told you what it had concluded. On Friday, at SXSW in Austin, the company decided to change that.
Gustav Söderström, Spotify’s co-CEO, announced Taste Profile: a new feature that surfaces the algorithmic model the platform has been building about each listener, and crucially lets users modify it directly. The beta will begin rolling out to Premium subscribers in New Zealand in the coming weeks.
The premise is straightforward enough. Taste Profile aggregates a listener’s behaviour across music, podcasts, and audiobooks into a single view: the genres explored recently, the artists listened to most, the patterns that define a listening day.
Advertisement
The 💜 of EU tech
The latest rumblings from the EU tech scene, a story from our wise ol’ founder Boris, and some questionable AI art. It’s free, every week, in your inbox. Sign up now!
Where a user notices the profile is wrong, too heavy on music they played years ago, or missing a phase they have been quietly working through, they can flag it. They can ask for more of a particular vibe, or less. They can describe a current context, training for an event, commuting on weekdays, and the system will factor that in when deciding what to surface on the Spotify homepage.
“This is the next step in our vision to make personalization more transparent, responsive, and truly yours,” Söderström told the SXSW audience.
Advertisement
Spotify cited an internal figure that more than 80% of its listeners name personalisation as what they value most about the service. The claim, which the company has referenced in various forms since at least 2023, positions algorithmic curation not just as a feature but as the primary reason people stay.
The competitive logic behind Taste Profile follows directly from that: if personalisation is the product, giving users more control over it is a way to deepen their investment in it.
The announcement comes roughly two months after Spotify expanded Prompted Playlist, a separate but related feature that lets users generate playlists by describing what they want in natural language, from its initial New Zealand testing to Premium users in the US and Canada in late January 2026, and subsequently to subscribers in Australia, Ireland, Sweden, and the UK in February. The sequencing is deliberate.
Both features push the same underlying argument: that the future of streaming personalisation is collaborative, not passive.
Advertisement
Where Prompted Playlist is generative, it creates something new from a description, Taste Profile is corrective. It works with the model that already exists, giving users a chance to audit and adjust what years of listening have written about them.
Whether someone has been an accidental customer of the algorithm (playing whatever appeared on the homepage, not particularly caring) or has strong views about the direction their recommendations have taken, the feature is designed to accommodate both. “You can shape your Taste Profile as much as you’d like,” the company said in its announcement, “or leave it and enjoy Spotify as usual.”
The beta will start in New Zealand, a market Spotify has used repeatedly for early-stage testing of AI-adjacent features, including the initial Prompted Playlist launch. No timeline was given for a broader global rollout. Taste Profile will be available to Premium subscribers only; there was no indication of when, or whether, it might reach free-tier accounts.
Spotify is marking 2026 as its 20th anniversary year, and its SXSW presence this week has been calibrated accordingly, concerts, a headline session with Söderström, country artist Lainey Wilson, and podcast host David Friedberg.
Advertisement
The Taste Profile announcement landed on the last day of the company’s main SXSW programming, providing a product note to accompany the celebration.
What the feature represents, beyond its functionality, is a shift in how Spotify frames its relationship with listeners. The algorithm has always existed; the company is now making the case that knowing it is there, and having some say in what it does, is a feature in itself.
Last June Texas Agriculture Commissioner Sid Miller said in a statement that Texans “have a God-given right to know what’s on their plate, and for millions of Texans, it better come from a pasture, not a lab. It’s plain cowboy logic that we must safeguard our real, authentic meat industry from synthetic alternatives.”
But California company Wildtype sells lab-grown salmon — and is suing Texas over its ban on cell-cultivated meat, the Austin Chronicle reported this week. The company’s founder says lab-grown salmon eliminates the mercury, microplastic, and antibiotic contamination commonly found in seafood. And one chef in Austin, Texas says lab-grown salmon is “awesome” and “something new”– at the only Texas restaurant that was serving it last summer:
Just two months after the salmon hit the menu, Texas banned the sale of cell-cultivated meat…
A lawsuit from Wildtype and one other FDA-approved cultivated meat company [argues] it’s anti-capitalism and unconstitutional… This law “was not enacted to protect the health and safety of Texas consumers — indeed, it allows the continued distribution of cultivated meat to consumers so long as it is not sold. Instead, SB 261 was enacted to stifle the growth of the cultivated meat industry to protect Texas’ conventional agricultural industry from innovative competition that is exclusively based outside of Texas….” [according to the lawsuit]. It was filed in September, immediately after the ban took effect, and cell-cultivated companies are awaiting judgment. That Texas ban would last two years, notes U.S. News and World Reports, adding that
Alabama, Florida, Indiana, Mississippi, Montana, and Nebraska have also passed bans, some temporary “on the manufacturing, sale or distribution of cell-cultured meat.” Meanwhile, a new five-year moratorium on lab-grown meat was signed this week by the governor of South Dakota “after rejecting a permanent ban last month,” reports South Dakota Searchlight:
The new law bars the sale, manufacture or distribution of “cell-cultured protein” products from July 1 this year through June 30, 2031. Violations are punishable by up to 30 days in jail, a fine of up to $500, or both. “But supporters of lab-grown meat are not going down without a fight,” adds U.S. News and World Reports, with another lawsuit also filed challenging a ban in Florida:
Advertisement
When Florida Gov. Ron DeSantis signed the ban in Florida, he described it as “fighting back against the global elite’s plan to force the world to eat meat grown in a petri dish or bugs to achieve their authoritarian goals.” He added that his administration “will save our beef.”
Amazon’s streaming service is getting a significant upgrade as the company transitions its ad-free tier to the new “Prime Video Ultra” branding. And with that branding, comes a price hike.
Prime Video Ultra
Launching April 10, the Ultra tier expands existing features, making it ideal for larger households. Beyond removing ads, the updated plan increases the concurrent stream limit to five devices, doubles the offline download capacity to 100 items, and gives users the option to watch in 4K/UHD. Amazon is also increasing the benefits for those who have the ad-supported video plan that comes gratis with a Prime membership. Those customers will be allowed to watch 4 concurrent streams, up from three, and download 50 items for offline viewing, up from 25. Continue Reading on AppleInsider | Discuss on our Forums
For years, third-party cybersecurity relationships between vendors and customers have relied on contracts and trust. That model is now showing its age. In the past year alone, 51% of UK organizations have reported a third party-related breach, while vendors have become ideal attack vectors for hostile actors.
Sam Kirkman
Director of EMEA Services at NetSPI.
Advertisement
At the same time, cyber attacks grow in their frequency and damage, even when they rely on relatively unsophisticated techniques. It is a clear shift: the way vendors reassure their customers of the strength of their security can no longer rely on words alone.
Trust based compliance to evidence-based security
What once worked for security vendors, trust-based compliance, has now become the bare minimum, as well as an outdated approach for modern cyber strategy and data protection.
Advertisement
Article continues below
Contracts and written assurances do little to protect organizations in practice, and too often, customers are left with limited insight into the real security posture of their vendors.
In the past few years, we have seen documentation, questionnaires and copious amounts of certifications which has come to overshadow demonstratable robustness. The emphasis has shifted towards ticking boxes, rather than proving strength.
Advertisement
Instead, we need to move from telling to showing; proof over promise.
An evidence-based model of security requires that vendors actively demonstrate that their security approach is measurably robust, measurable, and effective. Compliance does not equal resilience in today’s threat landscape, instead, only a consistent and proactive approach will do.
Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed!
Advertisement
Structural blindness
Of course most vendors are not deliberately hiding vulnerabilities from customers. The issues are latency and visibility. Point in-time assessments quickly become outdated and lose relevance as systems shifts, technology advances and new code is deployed.
A vendor deemed secure at the point of certification or contractual signing can carry material risks just weeks later without a consistent approach to vulnerability management.
Developing comprehensive visibility of vulnerabilities across an organization is often challenging. Unfortunately, some vendors choose a path of willful ignorance and blind optimism. This approach saves money for the vendor, at the expense of increasing the risk you take on as a customer.
Even when new vulnerabilities are found, customers often have little to no visibility. An ad hoc approach to third-party security has created a form of structural blindness where risk exists but remains unseen.
To address this, vendors must move towards continuously signaling operational and cyber resilience, rather than relying on static assurances.
Assurance in practice: penetration testing
In practical terms, this means on thing: continuous penetration testing.
Advertisement
For vendors performing infrequent or ad hoc tests, security teams struggle to keep up with the rapidly evolving landscape, leaving vulnerabilities unidentified and customers exposed.
By simulating real attacker behavior, vendors not only demonstrate their commitment to a strong security framework to customers, but it also actively improves their vulnerability management and reduces the very risk of a data breach in the first place.
Customers are assured with evidence; vendor’s security teams can sleep easy that their weaknesses have been addressed.
For organizations managing dozens, or hundreds, of third-party relationships, this level of visibility is critical to understanding where real risk resides and improving customer relationships.
Advertisement
It is time for CISOs to speak up
Supply chains have become prime targets for hostile actors, where data breaches lead to a domino effect of disruption across suppliers, warehouses and manufacturers. For instance, the devastating Jaguar Land Rover attack in September 2025 contributed to reducing real growth across the wider economy of the UK to just 0.1%.
It is critical that vendors begin to demonstrate, through evidence, that they are secure. CISOs are uniquely positioned to raise the bar and lead the charge in demanding third-party security teams are proving their robust cyber management.
To be clear, this is about a greater alignment between vendor and customer, not about punishing the vendors whose security might not be as strong as was hoped. Providing proof over promise represents a fundamental shift in the cybersecurity approach of both CISOs, third-parties and customer organizations.
Advertisement
Where CISOs are leading the charge, companies across all sectors can build up their resilience.
Words to live by
Cybersecurity can no longer rely on outdated and insufficient promises rooted in trust and contractual obligations.
The cyber landscape is in a constant state of evolution and change, and trust alone is no longer a reliable indicator of a mature security framework. Static assurances and point-in-time validations fail to reflect the realities of modern infrastructure, where risk evolves far faster than documentation ever can.
Advertisement
By embracing continuous penetration testing and empowering CISOs to demand that vendors demonstrably prove their security posture, organizations can fundamentally change how third-party risk is managed.
This shift moves the cybersecurity and business landscape away from blind trust that silently compromises data safety, and toward confidence grounded in ongoing, measurable assurance.
Proof over promises is an essential tenet of cybersecurity in the modern world.
This article was produced as part of TechRadarPro’s Expert Insights channel where we feature the best and brightest minds in the technology industry today. The views expressed here are those of the author and are not necessarily those of TechRadarPro or Future plc. If you are interested in contributing find out more here: https://www.techradar.com/news/submit-your-story-to-techradar-pro
I’ll admit that when I first started watching the recently released Blu-ray Disc of Peter Gabriel’s Taking the Pulse, it all felt very familiar. If you saw his orchestral tour around the time of Scratch My Back, or already own New Blood: Live in London on Blu-ray, a sense of déjà vu might set in. Look a little closer, however, and the differences start to emerge. What initially feels like a continuation of the same orchestral concept reveals a number of subtle but meaningful changes that aren’t immediately obvious.
In fact, the Blu-ray Disc packaging and frankly most of the press materials I had read, don’t offer much reason why fans should rush to add this show to their Peter Gabriel collection. Fortunately, I found more clues on Mr. Gabriel’s website, where his daughter Anna, who directed the film, offers some insight into her approach:
“When I spoke to my dad about shooting one of his shows I jumped at the chance of shooting at the Roman amphitheatre in Verona, Italy. Italians have always been very enthusiastic audience and give a lot back to the performer and to the camera. I made this film with my friend Andrew Gaston who was truly my collaborator throughout the process. It is always fun for me to shoot my father as I know the material and his performances so well I feel that I can capture a side to him that feels more personal. I also wanted to shoot the orchestra in an exciting way and along with Andrew’s editing I think we really captured the energy of the entire performance. I look forward to sharing this film.”
Indeed, Taking the Pulse offers a fresh visual perspective on this remarkable music when compared to New Blood: Live in London. Happily, it also delivers a different sonic experience, with a more immersive 5.1 surround sound presentation that better envelops the listener.
Advertisement
Presented in 24-bit/48 kHz resolution, the DTS-HD Master Audio soundtrack is generally quite rich. But my appreciation grew once I realized that the surround channels were being used for more than just crowd noise and venue ambience. One of my recurring complaints about many concert videos is that the surround mixes often feel lazy, failing to take full advantage of what the technology can actually deliver.
While most of the action resides in the front channels, the audio for Taking The Pulse nearly wraps around the listener. I imagine this perspective is bit more like what I’d like to think the orchestra conductor was hearing on stage or — perhaps better still — Peter Gabriel himself!
The sometimes dark, moody lighting looks great (expect lots of blues and reds as well as bursts of light and sparkle). In general, I think I preferred the intercuts and perspective offered in Taking The Pulse over New Blood Live In London (which, mind you, I’ve long enjoyed!).
All in all, Taking The Pulse is a winner Blu-ray video and for $18.29 on Amazon it seems to be an easy decision to pick this up if you are a fan.
Mark Smotroff is a deep music enthusiast / collector who has also worked in entertainment oriented marketing communications for decades supporting the likes of DTS, Sega and many others. He reviews vinyl for Analog Planet and has written for Audiophile Review, Sound+Vision, Mix, EQ, etc. You can learn more about him at LinkedIn.