Daily Tech Digest - June 10, 2025


Quote for the day:

"Life is not about finding yourself. Life is about creating yourself." -- Lolly Daskal


AI Is Making Cybercrime Quieter and Quicker

The rise of AI-enabled cybercrime is no longer theoretical. Nearly 72% of organisations In India said that they have encountered AI-powered cyber threats in the past year. These threats are scaling fast, with a 2X increase reported by 70% and a 3X increase by 12% of organisations. This new class of AI-powered threats are harder to detect and often exploit weaknesses in human behaviour, misconfigurations, and identity systems. In India, the top AI-driven threats reported include AI-assisted credential stuffing and brute force attacks, Deepfake impersonation in business email compromise (BEC), AI-powered malware (Polymorphic malware), Automated reconnaissance of attack surfaces, and AI-generated phishing emails. ... The most disruptive threats are no longer the most obvious. Topping the list are unpatched and zero-day exploits, followed closely by insider threats, cloud misconfigurations, software supply chain attacks, and human error. These threats are particularly damaging because they often go undetected by traditional defences, exploiting internal weaknesses and visibility gaps. As a result, these quieter, more complex risks are now viewed as more dangerous than well-known threats like ransomware or phishing. Traditional threats such as phishing and malware are still growing at a rate of ~10%, but this is comparatively modest —likely due to mature defences like endpoint protection and awareness training.


The Evolution and Future of the Relationship Between Business and IT

IT professionals increasingly serve as translators — converting executive goals into technical requirements, and turning technical realities into actionable business decisions. This fusion of roles has also led to the rise of cross-functional “fusion teams,” where IT and business units co-own projects from ideation through execution. ... Artificial Intelligence is already influencing how decisions are made and systems are managed. From intelligent automation to predictive analytics, AI is redefining productivity. According to a PwC report, AI is expected to contribute over $15 trillion to the global economy by 2030 — and IT organizations will play a pivotal role in enabling this transformation. At the same time, the lines between IT and the business will continue to blur. Platforms like low-code development tools, AI copilots, and intelligent data fabrics will empower business users to create solutions without traditional IT support — requiring IT teams to pivot further into governance, enablement, and strategy. Security, compliance, and data privacy will become even more important as businesses operate across fragmented and federated environments. ... The business-IT relationship has evolved from one rooted in infrastructure ownership to one centered on service integration, strategic alignment, and value delivery. IT is no longer just the department that runs servers or writes code — it’s the nervous system that connects capabilities, ensures reliability, and enables growth.


Can regulators trust black-box algorithms to enforce financial fairness?

Regulators, in their attempt to maintain oversight and comparability, often opt for rules-based regulation, said DiRollo. These are prescriptive, detailed requirements intended to eliminate ambiguity. However, this approach unintentionally creates a disproportionate burden on smaller institutions, he continued, DiRollo said, “Each bank must effectively build its own data architecture to interpret and implement regulatory requirements. For instance, calculating Risk-Weighted Assets (RWAs) demands banks to collate data across a myriad of systems, map this data into a bespoke regulatory model, apply overlays and assumptions to reflect the intent of the rule and interpret evolving guidance and submit reports accordingly.” ... Secondly around regulatory arbitrage. In this area, larger institutions with more sophisticated modelling capabilities can structure their portfolios or data in ways that reduce regulatory burdens without a corresponding reduction in actual risk. “The implication is stark: the fairness that regulators seek to enforce is undermined by the very framework designed to ensure it,” said DiRollo. While institutions pour effort into interpreting rules and submitting reports, the focus drifts from identifying and managing real risks. In practice, compliance becomes a proxy for safety – a dangerous assumption, in the words of DiRollo.


The legal questions to ask when your systems go dark

Legal should assume the worst and lean into their natural legal pessimism. There’s very little time to react, and it’s better to overreact than underreact (or not react at all). The legal context around cyber incidents is broad, but assume the worst-case scenario like a massive data breach. If that turns out to be wrong, even better! ... Even if your organization has a detailed incident response plan, chances are no one’s ever read it and that there will be people claiming “that’s not my job.” Don’t get caught up in that. Be the one who brings together management, IT, PR, and legal at the same table, and coordinate efforts from the legal perspective. ... If that means “my DPO will check the ROPA” – congrats! But if your processes are still a work in progress, you’re likely about to run a rapid, ad hoc data inventory: involving all departments, identifying data types, locations, and access controls. Yes, it will all be happening while systems are down and everyone’s panicking. But hey – serenity now, emotional damage later. You literally went to law school for this. ... You, as in-house or external legal support, really have to understand the organization and how its tech workflows actually function. I dream of a world where lawyers finally stop saying “we’ll just do the legal stuff,” because “legal stuff” remains abstract and therefore ineffective if you don’t put it in the context of a particular organization.


New Quantum Algorithm Factors Numbers With One Qubit

Ultimately, the new approach works because of how it encodes information. Classical computers use bits, which can take one of two values. Qubits, the quantum equivalent, can take on multiple values, because of the vagaries of quantum mechanics. But even qubits, once measured, can take on only one of two values, a 0 or a 1. But that’s not the only way to encode data in quantum devices, say Robert König and Lukas Brenner of the Technical University of Munich. Their work focuses on ways to encode information with continuous variables, meaning they can take on any values in a given range, instead of just certain ones. ... In the past, researchers have tried to improve on Shor’s algorithm for factoring by simulating a qubit using a continuous system, with its expanded set of possible values. But even if your system computes with continuous qubits, it will still need a lot of them to factor numbers, and it won’t necessarily go any faster. “We were wondering whether there’s a better way of using continuous variable systems,” König said. They decided to go back to basics. The secret to Shor’s algorithm is that it uses the number it’s factoring to generate what researchers call a periodic function, which has repeating values at regular intervals. Then it uses a mathematical tool called a quantum Fourier transform to identify the value of that period — how long it takes for the function to repeat.


What Are Large Action Models?

LAMs are LLMs trained on specific actions and enhanced with real connectivity to external data and systems. This makes the agents they power more robust than basic LLMs, which are limited to reasoning, retrieval and text generation. Whereas LLMs are more general-purpose, trained on a large data corpus, LAMs are more task-oriented. “LAMs fine-tune an LLM to specifically be good at recommending actions to complete a goal,” Jason Fournier, vice president of AI initiatives at the education platform Imagine Learning, told The New Stack. ... LAMs trained on internal actions could streamline industry-specific workflows as well. Imagine Learning, for instance, has developed a curriculum-informed AI framework to support teachers and students with AI-powered lesson planning. Fournier sees promise in automating administrative tasks like student registration, synthesizing data for educators and enhancing the learning experience. Or, Willson said, consider marketing: “You could tell an agentic AI platform with LAM technology, ‘Launch our new product campaign for the ACME software across all our channels with our standard messaging framework.'” Capabilities like this could save time, ensure brand consistency, and free teams to focus on high-level strategy.


Five mistakes companies make when retiring IT equipment: And how to avoid them

Outdated or unused IT assets often sit idle in storage closets, server rooms, or even employee homes for extended periods. This delay in decommissioning can create a host of problems. Unsecured, unused devices are prime targets for data breaches, theft, or accidental loss. Additionally, without a timely and consistent retirement process, organizations lose visibility into asset status, which can create confusion, non-compliance, or unnecessary costs. The best way to address this is by implementing in-house destruction solutions as an integrated part of the IT lifecycle. Rather than relying on external vendors or waiting until large volumes of devices pile up, organizations can equip themselves with high security data destruction machinery – such as hard drive shredders, degaussers, crushers, or disintegrators – designed to render data irretrievable on demand. This allows for immediate, on-site sanitization and physical destruction as soon as devices are decommissioned. Not only does this improve data control and reduce risk exposure, but it also simplifies chain-of-custody tracking by eliminating unnecessary handoffs. With in-house destruction capabilities, organizations can securely retire equipment at the pace their operations demand – no waiting, no outsourcing, and no compromise.


Event Sourcing Unpacked: The What, Why, and How

Event Sourcing offers significant benefits for systems that require persistent audit trails, rich debugging capabilities with event replay. It is especially effective in domains like finance, healthcare, e-commerce, and IoT, where every transaction or state change is critical and must be traceable. However, its complexity means that it isn’t ideal for every scenario. For applications that primarily engage in basic CRUD operations or demand immediate consistency, the overhead of managing an ever-growing event log, handling event schema evolution, and coping with eventual consistency can outweigh the benefits. In such cases, simpler persistence models may be more appropriate. When compared with related patterns, Event Sourcing naturally complements CQRS by decoupling read and write operations, and it enhances Domain-Driven Design by providing a historical record of domain events. Additionally, it underpins Event-Driven Architectures by facilitating loosely coupled, scalable communication. The decision to implement Event Sourcing should therefore balance its powerful capabilities against the operational and developmental complexities it introduces, ensuring it aligns with the project’s specific needs and long-term architectural goals.


Using Traffic Mirroring to Debug and Test Microservices in Production-Like Environments

At its core, traffic mirroring duplicates incoming requests so that, while one copy is served by the primary service, the other is sent to an identical service running in a test or staging environment. The response from the mirrored service is never returned to the client; it exists solely to let engineers observe, compare, or process data from real-world usage. ... Real-world traffic is messy. Certain bugs only appear when a request contains a specific sequence of API calls or unexpected data patterns. By mirroring production traffic to a shadow service, developers can catch these hard-to-reproduce errors in a controlled environment. ... Mirroring production traffic allows teams to observe how a new service version handles the same load as its predecessor. This testing is particularly useful for identifying regressions in response time or resource utilization. Teams can compare metrics like CPU usage, memory consumption, and request latency between the primary and shadow services to determine whether code changes negatively affect performance. Before rolling out a new feature, developers must ensure it works correctly under production conditions. Traffic mirroring lets a new microservice version be deployed with feature flags while still serving requests from the stable version.


Don’t be a victim of high cloud costs

The simplest reason for the rising expenses associated with cloud services is that major cloud service providers consistently increase their prices. Although competition among these providers helps keep prices stable to some extent, businesses now face inflation, the introduction of new premium services, and the complex nature of pricing models, which are often shrouded in mystery. All these factors complicate cost management. Meanwhile, many businesses have inefficient usage patterns. The typical approach to adoption involves migrating existing systems to the cloud without modifying or improving their functions for cloud environments. This “lift and shift” shortcut often leads to inefficient resource allocation and unnecessary expenses. ... First, before embracing cloud technology for its advantages, companies should develop a well-defined plan that outlines the rationale, objectives, and approach to using cloud services. Identify which tasks are suitable for cloud deployment and which are not, and assess whether a public, private, or hybrid cloud setup aligns with your business and budget objectives. Second, before transferring data, ensure that you optimize your tasks to improve efficiency and performance. Please resist the urge to move existing systems to the cloud in their current state. ... Third, effectively managing cloud expenses relies on implementing strong governance practices.

No comments:

Post a Comment