Daily Tech Digest - April 07, 2025


Quote for the day:

"Failure isn't fatal, but failure to change might be" -- John Wooden



How enterprise IT can protect itself from genAI unreliability

The AI-watching-AI approach is scarier, although a lot of enterprises are giving it a go. Some are looking to push any liability down the road by partnering with others to do their genAI calculations for them. Still others are looking to pay third-parties to come in and try and improve their genAI accuracy. The phrase “throwing good money after bad” immediately comes to mind. The lack of effective ways to improve genAI reliability internally is a key factor in why so many proof-of-concept trials got approved quickly, but never moved into production. Some version of throwing more humans into the mix to keep an eye on genAI outputs seems to be winning the argument, for now. “You have to have a human babysitter on it. AI watching AI is guaranteed to fail,” said Missy Cummings, a George Mason University professor and director of Mason’s Autonomy and Robotics Center (MARC). “People are going to do it because they want to believe in the (technology’s) promises. People can be taken in by the self-confidence of a genAI system,” she said, comparing it to the experience of driving autonomous vehicles (AVs). When driving an AV, “the AI is pretty good and it can work. But if you quit paying attention for a quick second,” disaster can strike, Cummings said. “The bigger problem is that people develop an unhealthy complacency.”


Why neglecting AI ethics is such risky business - and how to do AI right

The struggle often comes from the lack of a common vocabulary around AI. This is why the first step is to set up a cross-organizational strategy that brings together technical teams as well as legal and HR teams. AI is transformational and requires a corporate approach. Second, organizations need to understand what the key tenets of their AI approach are. This goes beyond the law and encompasses the values they want to uphold. Third, they can develop a risk taxonomy based on the risks they foresee. Risks are based on legal alignment, security, and the impact on the workforce. ... As a starting point, enterprises will need to establish clear policies, principles, and guidelines on the sustainable use of AI. This creates a baseline for decisions around AI innovation and enables teams to make the right choices around the type of AI infrastructure, models, and algorithms they will adopt. Additionally, enterprises need to establish systems to effectively track, measure, and monitor environmental impact from AI usage and demand this from their service providers. We have worked with clients to evaluate current AI policies, engage internal and external stakeholders, and develop new principles around AI and the environment before training and educating employees across several functions to embed thinking in everyday processes.


The risks of entry-level developers over relying on AI

Some CISOs are concerned about the growing reliance on AI code generators — especially among junior developers — while others take a more relaxed, wait-and-see approach, saying that this might be an issue in the future rather than an immediate threat. Karl Mattson, CISO at Endor Labs, argues that the adoption of AI is still in its early stages in most large enterprises and that the benefits of experimentation still outweigh the risks. ... Tuskira’s CISO lists two major issues: first, that AI-generated security code may not be hardened against evolving attack techniques; and second, that it may fail to reflect the specific security landscape and needs of the organization. Additionally, AI-generated code might give a false sense of security, as developers, particularly inexperienced ones, often assume it is secure by default. Furthermore, there are risks associated with compliance and violations of licensing terms or regulatory standards, which can lead to legal issues down the line. “Many AI tools, especially those generating code based on open-source codebases, can inadvertently introduce unvetted, improperly licensed, or even malicious code into your system,” O’Brien says. Open-source licenses, for example, often have specific requirements regarding attribution, redistribution, and modifications, and relying on AI-generated code could mean accidentally violating these licenses.


Language models in generative AI – does size matter?

Firstly, using SLMs rather than full-blown LLMs can bring the cost of that multi-agent system down considerably. Employing smaller and more lightweight language models to fulfill specific requirements will be more cost-effective than using LLMs for every step in an agentic AI system. This approach involves looking at what would be the right component for each element of a multi-agent system, rather than automatically thinking that a “best of breed” approach is the best approach. Secondly, using agentic AI for generative AI use cases should be adopted where multi-agent processes can provide more value per transaction than simpler single-agent models. The choice here affects how you think about pricing your service, what customers expect from AI and how you will deliver your service overall. Alongside looking at the technical and architecture elements for AI, you will also have to consider what your line of business team wants to achieve. While simple AI agents can carry out specific tasks or automate repetitive tasks, they generally require human input to complete those requests. Where agentic AI takes things further is through delivering greater autonomy within business processes through employing that multi-agent approach to constantly adapt to dynamic environments. With agentic AI, companies can use AI to independently create, execute and optimize results around that business process workflow. 


Lessons from a Decade of Complexity: Microservices to Simplicity

This shift made us stop and think: if fast growth isn’t the priority anymore, is microservices still the right choice? ... After going through years of building and maintaining systems with microservices, we’ve learned a lot, especially about what really matters in choosing an architecture. Here are some key takeaways that guide how we think about system design today: Be pragmatic, not idealistic: Don’t get caught up in trendy architecture patterns just because they sound impressive. Focus on what makes sense for your team and your situation. Not every new system needs to start with microservices, especially if the problems they solve aren’t even there yet. Start simple: The simplest solution is often the best one. It’s easier to build, easier to understand, and easier to change. Keeping things simple takes discipline, but it saves time and pain in the long run. Split only when it really makes sense: Don’t break things apart just because “that’s what we do”. Split services when there’s a clear technical reason, like performance, resource needs, or special hardware. Microservices are just a tool: They’re not good or bad by themselves. What matters is whether they help your team move faster, stay flexible, and solve real problems. Every choice comes with tradeoffs: No architecture is perfect. Every decision has upsides and downsides. What’s important is to be aware of those tradeoffs and make the best call for your team.


Massive modernization: Tips for overhauling IT at scale

A core part of digital transformation is decommissioning legacy apps, upgrading aging systems, and modernizing the tech stack. Yet, as appealing as it is for employees to be able to use modern technologies, decommissioning and replacing systems is arduous for IT. ... “You almost do what I call putting lipstick on a pig, which is modernizing your legacy ecosystem with wrappers, whether it be web wrappers, front end and other technologies that allow customers to be able to interact with more modern interfaces,” he says. ... When an organization is truly legacy, most will likely have very little documentation of how those systems can be supported, Mehta says. That was the case for National Life, and it became the first roadblock. “You don’t know what you don’t know until we begin,” he says. This is where the archaeological dig metaphor comes in. “You’re building a new city over the top of the old city, but you’ve got to be able to dig it only enough so you don’t collapse the foundation.” IT has to figure out everything a system touches, “because over time, people have done all kinds of things to it that are not clearly documented,” Mehta says. ... “You have to have a plan to get rid of” legacy systems. He also discovered that “decommissioning is not free. Everybody thinks you just shut a switch off and legacy systems are gone. Legacy decommissioning comes at a cost. You have to be willing to absorb that cost as part of your new system. That was a lesson learned; you cannot ignore that,” he says.


Culture is not static: Prasad Menon on building a thriving workplace at Unplugged 3

To cultivate a thriving workplace, organisations must engage in active listening. Employees should have structured platforms to voice their concerns, aspirations, and feedback without hesitation. At Amagi, this commitment to deep listening is reinforced by technology. The company has implemented an AI-powered chatbot named Samb, which acts as a "listening manager," facilitating real-time employee feedback collection. This tool ensures that concerns and suggestions are acknowledged and addressed within 15 days, allowing for a more responsive and agile work environment. "Culture is not just a feel-good factor—it must be measured and linked to results," Menon emphasised. To track and optimise cultural impact, Amagi has developed a "happiness index" that measures employee well-being across financial, mental, and physical dimensions. By using data to evaluate cultural effectiveness, the organisation ensures that workplace culture is not just an abstract ideal but a tangible force driving business success. ... At the core of Amagi’s culture is a commitment to becoming "the happiest workplace in the world." This vision is driven by a leadership model that prioritises genuine care, consistency, and empowerment. Leaders at Amagi undergo a six-month cultural immersion programme designed to equip them with the skills needed to foster a safe, inclusive, and high-performing work environment.


Speaking the Board’s Language: A CISO’s Guide to Securing Cybersecurity Budget

A major challenge for CISOs in budget discussions is making cybersecurity risk feel tangible. Cyber risks often remain invisible – that is, until a breach happens. Traditional tools like heat maps, which visually represent risk by color-coding potential threats, can be misleading or oversimplified. While they offer a high-level view of risk areas, heat maps fail to provide a concrete understanding of the actual financial impact of those risks. This makes it essential to shift from qualitative risk assessments like heat maps to cyber risk quantification (CRQ), which assigns a measurable dollar value to potential threats and mitigation efforts. ... The biggest challenge CISOs face isn’t just securing budget – it’s making sure decision-makers understand why they need it. Boards and executives don’t think in terms of firewalls and threat detection; they care about business continuity, revenue protection and return on investment (ROI). For cyber investments, though, ROI is not typically the figure security experts turn to to validate these investments, largely because of the difficulties in estimating the value of risk reduction. However, new approaches to cyber risk quantification have made this a reality. With models validated by real-world loss data, it is now possible to produce an ROI figure. 


Can AI predict who will commit crime?

Simulating the conditions for individual offending is not the same as calculating the likelihood of storms or energy outages. Offending is often situational and is heavily influenced by emotional, psychological and environmental elements (a bit like sport – ever wondered why Predictive AI hasn’t put bookmakers out of business yet?). Sociological factors also play a big part in rehabilitation which, in turn, affects future offending. Predictive profiling relies on past behaviour being a good indicator of future conduct. Is this a fair assumption? Occupational psychologists say past behaviour is a reliable predictor of future performance – which is why they design job selection around it. Unlike financial instruments which warn against assuming future returns from past rewards, human behaviour does have a perennial quality. Leopards and spots come to mind. ... Even if the data could reliably tell us who will be charged with, prosecuted for and convicted of which specific offence in the future, what should the police do about it now? Implant a biometric chip and have them under perpetual surveillance to stop them doing what they probably didn’t know they were going to do? Fine or imprison them? (how much, for how long?). What standard of proof will the AI apply to its predictions? Beyond a reasonable doubt? How will we measure the accuracy of the process? 


CISOs battle security platform fatigue

“Adopting more security tools doesn’t guarantee better cybersecurity,” says Jonathan Gill, CEO at Panaseer. “These tools can only report on what they can see – but they don’t know what they’re missing.” This fragmented visibility leaves security leaders making high-stakes decisions based on partial information. Without a verified, comprehensive system of record for all assets and security controls, many organizations are operating under what Gill calls an “illusion of visibility.” “Without a true denominator,” he explains, “CISOs are unable to confidently assess coverage gaps or prove compliance with evolving regulatory demands.” And those blind spots aren’t just theoretical. Every overlooked asset or misconfigured control becomes an open door for attackers — and they’re getting better at finding them. “Each of these coverage gaps represents risk,” Gill warns, “and they are increasingly easy for attackers to find and exploit.” The lack of clear visibility also muddies accountability. “This creates dark corners that go overlooked – servers and applications are left without owners, making it hard to assign responsibility for fixing issues,” Gill says. Even when gaps are known, security teams often find themselves drowning in data from too many tools, struggling to separate signal from noise. 

No comments:

Post a Comment