Showing posts with label AiTM. Show all posts
Showing posts with label AiTM. Show all posts

Daily Tech Digest - April 28, 2026


Quote for the day:

"Authentic leaders give credit when and where it is due." -- Samuel Adams


🎧 Listen to this digest on YouTube Music

▶ Play Audio Digest

Duration: 23 mins • Perfect for listening on the go.


Zero trust at scale: Practical strategies for global enterprises

In the article "Zero Trust at Scale: Practical Strategies for Global Enterprises," Shibu Paul of Array Networks highlights the necessity of Zero Trust Architecture (ZTA) as traditional perimeter-based security fails against modern, decentralized cyber threats. Built on the core principle of "never trust, always verify," ZTA replaces outdated assumptions of internal safety with rigorous, continuous authentication for every user and device. The framework relies on four critical pillars: continuous verification, least-privilege access, micro-segmentation, and real-time monitoring. Paul notes that while 86% of organizations have begun their Zero Trust journey, only 2% have fully matured their implementation. Practical strategies for global deployment include robust Identity and Access Management (IAM), multi-factor authentication, and sophisticated data loss prevention (DLP) across cloud and mobile environments. Despite integration complexities and the need for a significant cultural shift, the benefits are quantifiable; organizations adopting ZTA report a decrease in security incidents from an average of 18.2 to 8.5 per month and a 50% reduction in incident response times. Ultimately, Paul argues that Zero Trust is no longer an optional competitive advantage but a fundamental requirement for maintaining operational resilience and securing sensitive data within the increasingly complex digital landscape of contemporary global enterprises.


Slow down to speed up: Why steadfast IT leadership is critical in the age of AI

In the CIO.com article, "Slow down to speed up: Why steadfast IT leadership is critical in the age of AI," author Glen Brookman argues that while the pressure to adopt artificial intelligence is immense, sustainable success requires a "readiness-first" approach rather than raw speed. Brookman asserts that AI acts as an amplifier; it strengthens robust foundations but ruthlessly exposes weaknesses in data governance, security, and infrastructure. The core philosophy of "slowing down to speed up" suggests that leaders must prioritize the hard work of preparation—cleaning data sets, upgrading legacy systems, and establishing rigorous governance—to ensure innovation can take root. He warns that moving too quickly creates a "gravity doesn’t exist" mindset, where organizations believe AI can paper over process gaps, ultimately leading to fragility and risk. Brookman highlights that 75 percent of Canadian organizations utilize structured pilots to maintain discipline and avoid scattered experimentation. Ultimately, the CIO’s role is not to obstruct progress but to provide the "engine and steering" necessary for safe acceleration. By leading with clarity and technical rigor, IT executives ensure that their organizations are not just the first to deploy AI, but the most prepared to win in the long term.


Stopping AiTM attacks: The defenses that actually work after authentication succeeds

Adversary-in-the-Middle (AiTM) attacks have fundamentally shifted the cybersecurity landscape by bypassing traditional multi-factor authentication (MFA) through the real-time interception of session tokens. While many organizations respond to these threats by strengthening the authentication layer with FIDO2 or passkeys—which are effective at preventing initial credential theft—this approach is often incomplete because it fails to address what happens after a session is established. Since session cookies typically act as "bearer tokens" that are not cryptographically bound to a specific device, an attacker who captures one can impersonate a user without further challenges. Effective defense requires moving beyond the login event to implement post-authentication controls. Key strategies include session binding, which links a token to a specific hardware context, and continuous behavioral monitoring to detect anomalies like "impossible travel" or unusual API activity. Additionally, organizations should enforce strict conditional access policies that evaluate device posture and location in real time. Reducing token lifetimes and implementing rapid revocation capabilities for both access and refresh tokens are also critical for minimizing an attacker's window of opportunity. Ultimately, the article argues that security teams must treat "successful MFA" as a starting point for monitoring rather than an absolute guarantee of trust.


Deepfake Voice Attacks are Outpacing Defenses: What Security Leaders Should Know

"Deepfake Voice Attacks are Outpacing Defenses" by Marshall Bennett highlights the alarming rise of AI-generated audio and video fraud, which surged by 680% in 2025. The article warns that attackers need only three seconds of a person's voice—often harvested from social media or public appearances—to create a convincing, real-time replica. These sophisticated deepfakes are increasingly used to bypass traditional security stacks by targeting the human element, specifically finance and HR teams. High-profile incidents, such as a $25.6 million theft from the firm Arup and a $499,000 fraud in Singapore, illustrate the devastating financial impact of these "thin slice" attacks. Beyond financial theft, AI personas are even infiltrating hiring pipelines to gain internal system access. Because modern security software is often blind to conversational fraud, Bennett argues that the most effective defense is building human intuition. He recommends that organizations implement strict verification protocols, such as verbal passcodes and mandatory callbacks for high-value transfers. Ultimately, security leaders must move beyond annual compliance training to active simulations that build a "reflex to pause," ensuring employees can recognize and verify urgent requests before falling victim to a synthetic voice.


How AI is Changing Programming Language Usage

The article "How AI Is Changing Programming Language Usage" explores the profound impact of generative AI and Large Language Models (LLMs) on the software development landscape. As AI-powered tools like GitHub Copilot and ChatGPT become integral to the coding process, they are fundamentally altering which programming languages developers prioritize and how they interact with them. Python continues to dominate due to its extensive libraries and its role as the primary language for AI development itself. However, the rise of AI is also revitalizing interest in lower-level languages like Rust and C++, which are essential for building the high-performance infrastructure that powers AI models. Furthermore, the article highlights a shift in the "barrier to entry" for coding; natural language is increasingly becoming a bridge, allowing non-experts to generate functional code in diverse languages. This democratization suggests a future where the specific syntax of a language may matter less than a developer’s ability to architect systems and provide precise prompts. While AI enhances productivity by automating boilerplate tasks, it also introduces risks, such as the propagation of legacy bugs or "hallucinated" code, requiring developers to evolve into more critical reviewers and system designers rather than just manual coders.


Short-Lived Credentials in Agentic Systems: A Practical Trade-off Guide

In the article "Short-Lived Credentials in Agentic Systems: A Practical Trade-off Guide," Dwayne McDaniel highlights the critical role of short-lived credentials as a foundational security control for autonomous AI agents. As these systems transition from theoretical designs to production environments, they interact with numerous APIs, data stores, and cloud resources, significantly expanding the potential attack surface. Because agents can improvise and operate autonomously, long-lived "standing permissions" represent a major risk; if leaked, they allow for extended periods of unauthorized access and lateral movement. McDaniel argues that a mature security posture requires tying credential lifetimes—or Time to Live (TTL)—directly to the agent’s specific task, privilege level, and execution model. For instance, user-facing copilots might utilize a 5-to-15-minute TTL, whereas complex orchestration workflows require segmented access rather than a single broad token. By implementing a system where a broker or vault issues scoped, ephemeral credentials only after verifying the workload’s identity, organizations can drastically reduce the "blast radius" of a leak. Ultimately, while short-lived credentials increase operational complexity, they are essential for ensuring that autonomous agents remain accountable, revocable, and secure within modern digital ecosystems.


AI regulation set to become US midterm battleground

As the 2026 U.S. midterm elections approach, artificial intelligence regulation has emerged as a high-stakes political battleground, fueled by record-breaking campaign spending and a sharp ideological divide. Pro-innovation groups, such as Leading the Future and Innovation Council Action, have amassed over $225 million to support candidates favoring a "light-touch" regulatory approach, arguing that strict guardrails would stifle American competitiveness against China. These organizations are largely backed by tech industry leaders and align with a federal push to preempt state-level regulations. Conversely, groups like Public First Action, supported by Anthropic, are mobilizing tens of millions to advocate for robust safety measures to protect workers and families from AI risks. This clash is intensified by a volatile regulatory environment where the White House’s National AI Policy Framework faces significant pushback from states like California and Colorado, which have enacted their own stringent transparency and consumer protection laws. With polls indicating that a majority of Americans favor stronger oversight, the debate over whether to centralize authority or allow a patchwork of state rules has become a defining issue for voters. Consequently, the midterm results will likely determine the trajectory of U.S. technological governance for years to come.


3 Ways To Turn Your Leadership Gaps Into Your Purpose-Driven Advantage

In her Forbes article, "3 Ways To Turn Your Leadership Gaps Into Your Purpose-Driven Advantage," Luciana Paulise argues that leadership flaws are not mere liabilities but essential catalysts for professional growth and organizational impact. She asserts that the traditional "superhero" leadership model is increasingly obsolete in a modern workforce that prioritizes authenticity and shared values. Paulise outlines a transformative framework where leaders first practice radical self-awareness by identifying their specific "gaps"—whether in technical skills or emotional intelligence—and reframing them as opportunities for team collaboration. By openly acknowledging these limitations, leaders foster a culture of psychological safety that encourages others to step up and fill those voids, thereby creating a more resilient, distributed leadership structure. The article emphasizes that purpose-driven leadership emerges when personal vulnerabilities align with the organization’s mission, allowing for more genuine connections with employees. Paulise concludes that by leaning into their imperfections, executives can build higher levels of trust and engagement, shifting the focus from individual performance to collective achievement. This approach not only bridges capability gaps but also turns them into a strategic advantage that drives long-term retention and social impact.


Trying Pair Programming With An LLM Chatbot

The article "Trying Pair Programming With An LLM Chatbot" on Hackaday explores the potential of Large Language Models (LLMs) as coding partners, framed through the lens of an introverted developer who typically avoids the social friction of traditional pair programming. The author, skeptical of the hype surrounding "vibe coding," conducts an experiment using GitHub Copilot to see if an AI assistant can provide the benefits of collaboration without the awkwardness of human interaction. The narrative details a technical journey involving the STM32 microcontroller and the challenges of digging through complex datasheets and reference manuals. Unfortunately, the experience is marred by technical instability, such as the Copilot chat failing to load, and the realization that unlike human partners, AI can become abruptly unresponsive. Ultimately, the piece highlights a growing divide in the developer community: while some see LLMs as a "universal API" for specialized tasks like sentiment analysis, others warn that delegating engineering to statistical models can degrade critical thinking and lead to "AI slop." The experiment serves as a cautionary tale about model selection and the limitations of current AI tools in high-stakes, "close-to-the-metal" programming environments.


Your IAM was built for humans, AI agents don’t care

The Help Net Security article "Your IAM was built for humans, AI agents don't care" argues that traditional Identity and Access Management (IAM) systems are fundamentally ill-equipped for the rise of autonomous AI agents. While modern IT environments are increasingly dominated by non-human identities—accounting for over 90% of authentications—most IAM architectures still rely on the "single-gate" assumption: once a user is authenticated, they are trusted throughout a multi-step workflow. This creates a structural vulnerability when AI agents act on behalf of users, often utilizing broad, pre-provisioned permissions that lack visibility and granular control. The author warns against the industry's instinct to treat agents like employees by applying directory-based lifecycle management, which leads to "identity sprawl" as agents spawn and dissolve in seconds. Instead, the piece advocates for a shift toward runtime authorization where access tokens serve as carriers of dynamic context—defining who the agent represents and exactly what task it is authorized to perform at that specific moment. By transitioning from static credentials to just-in-time, task-scoped authorization, organizations can close the security gap in API chains and ensure that permissions disappear the moment a task is completed, effectively mitigating the risks of standing access.

Daily Tech Digest - April 24, 2025


Quote for the day:

“Remember, teamwork begins by building trust. And the only way to do that is to overcome our need for invulnerability.” -- Patrick Lencioni



Algorithm can make AI responses increasingly reliable with less computational overhead

The algorithm uses the structure according to which the language information is organized in the AI's large language model (LLM) to find related information. The models divide the language information in their training data into word parts. The semantic and syntactic relationships between the word parts are then arranged as connecting arrows—known in the field as vectors—in a multidimensional space. The dimensions of space, which can number in the thousands, arise from the relationship parameters that the LLM independently identifies during training using the general data. ... Relational arrows pointing in the same direction in this vector space indicate a strong correlation. The larger the angle between two vectors, the less two units of information relate to one another. The SIFT algorithm developed by ETH researchers now uses the direction of the relationship vector of the input query (prompt) to identify those information relationships that are closely related to the question but at the same time complement each other in terms of content. ... By contrast, the most common method used to date for selecting the information suitable for the answer, known as the nearest neighbor method, tends to accumulate redundant information that is widely available. The difference between the two methods becomes clear when looking at an example of a query prompt that is composed of several pieces of information.


Bring Your Own Malware: ransomware innovates again

The approach taken by DragonForce and Anubis shows that cybercriminals are becoming increasingly sophisticated in the way they market their services to potential affiliates. This marketing approach, in which DragonForce positions itself as a fully-fledged service platform and Anubis offers different revenue models, reflects how ransomware operators behave like “real” companies. Recent research has also shown that some cybercriminals even hire pentesters to test their ransomware for vulnerabilities before deploying it. So it’s not just dark web sites or a division of tasks, but a real ecosystem of clear options for “consumers.” We may also see a modernization of dark web forums, which currently resemble the online platforms of the 2000s. ... Although these developments in the ransomware landscape are worrying, Secureworks researchers also offer practical advice for organizations to protect themselves. Above all, defenders must take “proactive preventive” action. Fortunately and unfortunately, this mainly involves basic measures. Fortunately, because the policies to be implemented are manageable; unfortunately, because there is still a lack of universal awareness of such security practices. In addition, organizations must develop and regularly test an incident response plan to quickly remediate ransomware activities.


Phishing attacks thrive on human behaviour, not lack of skill

Phishing draws heavily from principles of psychology and classic social engineering. Attacks often play on authority bias, prompting individuals to comply with requests from supposed authority figures, such as IT personnel, management, or established brands. Additionally, attackers exploit urgency and scarcity by sending warnings of account suspensions or missed payments, and manipulate familiarity by referencing known organisations or colleagues. Psychologs has explained that many phishing techniques bear resemblance to those used by traditional confidence tricksters. These attacks depend on inducing quick, emotionally-driven decisions that can bypass normal critical thinking defences. The sophistication of phishing is furthered by increasing use of data-driven tactics. As highlighted by TechSplicer, attackers are now gathering publicly available information from sources like LinkedIn and company websites to make their phishing attempts appear more credible and tailored to the recipient. Even experienced professionals often fall for phishing attacks, not due to a lack of intelligence, but because high workload, multitasking, or emotional pressure make it difficult to properly scrutinise every communication. 

What Steve Jobs can teach us about rebranding

Humans like to think of themselves as rational animals, but it comes as no news to marketers that we are motivated to a greater extent by emotions. Logic brings us to conclusions; emotion brings us to action. Whether we are creating a poem or a new brand name, we won’t get very far if we treat the task as an engineering exercise. True, names are formed by putting together parts, just as poems are put together with rhythmic patterns and with rhyming lines, but that totally misses what is essential to a name’s success or a poem’s success. Consider Microsoft and Apple as names. One is far more mechanical, and the other much more effective at creating the beginning of an experience. While both companies are tremendously successful, there is no question that Apple has the stronger, more emotional experience. ... Different stakeholders care about different things. Employees need inspiration; investors need confidence; customers need clarity on what’s in it for them. Break down these audiences and craft tailored messages for each group. Identifying the audience groups can be challenging. While the first layer is obvious—customers, employees, investors, and analysts—all these audiences are easy to find and message. However, what is often overlooked is the individuals in those audiences who can more positively influence the rebrand. It may be a particular journalist, or a few select employees. 


Coaching AI agents: Why your next security hire might be an algorithm

Like any new team member, AI agents need onboarding before operating at maximum efficacy. Without proper onboarding, they risk misclassifying threats, generating excessive false positives, or failing to recognize subtle attack patterns. That’s why more mature agentic AI systems will ask for access to internal documentation, historical incident logs, or chat histories so the system can study them and adapt to the organization. Historical security incidents, environmental details, and incident response playbooks serve as training material, helping it recognize threats within an organization’s unique security landscape. Alternatively, these details can help the agentic system recognize benign activity. For example, once the system knows what are allowed VPN services or which users are authorized to conduct security testing, it will know to mark some alerts related to those services or activities as benign. ... Adapting AI isn’t a one-time event, it’s an ongoing process. Like any team member, agentic AI deployments improve through experience, feedback, and continuous refinement. The first step is maintaining human-in-the-loop oversight. Like any responsible manager, security analysts must regularly review AI-generated reports, verify key findings, and refine conclusions when necessary. 


Cyber insurance is no longer optional, it’s a strategic necessity

Once the DPDPA fully comes into effect, it will significantly alter how companies approach data protection. Many enterprises are already making efforts to manage their exposure, but despite their best intentions, they can still fall victim to breaches. We anticipate that the implementation of DPDPA will likely lead to an increase in the uptake of cyber insurance. This is because the Act clearly outlines that companies may face penalties in the event of a data breach originating from their environment. Since cyber insurance policies often include coverage for fines and penalties, this will become an increasingly important risk-transfer tool. ... The critical question has always been: how can we accurately quantify risk exposure? Specifically, if a certain event were to occur, what would be the financial impact? Today, there are advanced tools and probabilistic models available that allow organisations to answer this question with greater precision. Scenario analyses can now be conducted to simulate potential events and estimate the resulting financial impact. This, in turn, helps enterprises determine the appropriate level of insurance coverage, making the process far more data-driven and objective. Post-incident technology also plays a crucial role in forensic analysis. When an incident occurs, the immediate focus is on containment. 


Adversary-in-the-Middle Attacks Persist – Strategies to Lessen the Impact

One of the most recent examples of an AiTM attack is the attack on Microsoft 365 with the PhaaS toolkit Rockstar 2FA, an updated version of the DadSec/Phoenix kit. In 2024, a Microsoft employee accessed an attachment that led them to a phony website where they authenticated the attacker’s identity through the link. In this instance, the employee was tricked into performing an identity verification session, which granted the attacker entry to their account. ... As more businesses move online, from banks to critical services, fraudsters are more tempted by new targets. The challenges often depend on location and sector, but one thing is clear: Fraud operates without limitations. In the United States, AiTM fraud is progressively targeting financial services, e-commerce and iGaming. For financial services, this means that cybercriminals are intercepting transactions or altering payment details, inducing hefty losses. Concerning e-commerce and marketplaces, attackers are exploiting vulnerabilities to intercept and modify transactions through data manipulation, redirecting payments to their accounts. ... As technology advances and fraud continues to evolve with it, we face the persistent challenge of increased fraudster sophistication, threatening businesses of all sizes. 


From legacy to lakehouse: Centralizing insurance data with Delta Lake

Centralizing data and creating a Delta Lakehouse architecture significantly enhances AI model training and performance, yielding more accurate insights and predictive capabilities. The time-travel functionality of the delta format enables AI systems to access historical data versions for training and testing purposes. A critical consideration emerges regarding enterprise AI platform implementation. Modern AI models, particularly large language models, frequently require real-time data processing capabilities. The machine learning models would target and solve for one use case, but Gen AI has the capability to learn and address multiple use cases at scale. In this context, Delta Lake effectively manages these diverse data requirements, providing a unified data platform for enterprise GenAI initiatives. ... This unification of data engineering, data science and business intelligence workflows contrasts sharply with traditional approaches that required cumbersome data movement between disparate systems (e.g., data lake for exploration, data warehouse for BI, separate ML platforms). Lakehouse creates a synergistic ecosystem, dramatically accelerating the path from raw data collection to deployed AI models generating tangible business value, such as reduced fraud losses, faster claims settlements, more accurate pricing and enhanced customer relationships.


How AI and Data-Driven Decision Making Are Reshaping IT Ops

Rather than relying on intuition, IT decision-makers now lean on insights drawn from operational data, customer feedback, infrastructure performance, and market trends. The objective is simple: make informed decisions that align with broader business goals while minimizing risk and maximizing operational efficiency. With the help of analytics platforms and business intelligence tools, these insights are often transformed into interactive dashboards and visual reports, giving IT teams real-time visibility into performance metrics, system anomalies, and predictive outcomes. A key evolution in this approach is the use of predictive intelligence. Traditional project and service management often fall short when it comes to anticipating issues or forecasting success. ... AI also helps IT teams uncover patterns that are not immediately visible to the human eye. Predictive models built on historical performance data allow organizations to forecast demand, manage workloads more efficiently, and preemptively resolve issues before they disrupt service. This shift not only reduces downtime but also frees up resources to drive innovation across the enterprise. Moreover, companies that embrace data as a core business asset tend to nurture a culture of curiosity and informed experimentation. 


The DFIR Investigative Mindset: Brett Shavers On Thinking Like A Detective

You must be technical. You have to be technically proficient. You have to be able to do the actual technical work. And I’m not to rely on- not to bash a vendor training for a tool training, you have to have tool training, but you have to have exact training on “This is what the registry is, this is how you pull the-” you have to have that information first. The basics. You gotta have the basics, you have the fundamentals. And a lot of people wanna skip that. ... The DF guys, it’s like a criminal case. It’s “This is the computer that was in the back of the trunk of a car, and that’s what we got.” And the IR side is “This is our system and we set up everything and we can capture what we want. We can ignore what we want.” So if you’re looking at it like “Just in case something is gonna be criminal we might want to prepare a little bit,” right? So that makes DF guys really happy. If they’re coming in after the fact of an IR that becomes a case, a criminal case or a civil litigation where the DF comes in, they go, “Wow, this is nice. You guys have everything preserved, set up as if from the start you were prepared for this.” And it’s “We weren’t really prepared. We were prepared for it, we’re hoping it didn’t happen, we got it.” But I’ve walked in where drives are being wiped on a legal case.