Showing posts with label surveillance. Show all posts
Showing posts with label surveillance. Show all posts

Daily Tech Digest - February 06, 2026


Quote for the day:

"When you say my team is no good, all I hear is that I failed as a leader." -- Gordon Tredgold



Everyone works with AI agents, but who controls the agents?

Over the past year, there has been a lot of talk about MCP and A2A, protocols that allow agents to communicate with each other. But more and more agents that are now becoming available support and use them. Agents will soon be able to easily exchange information and transfer tasks to each other to achieve much better results. Currently, 50 percent of AI agents in organizations still work as a silo. This means that no context or data from external systems is added. The need for context is now clear to many organizations. 96 percent of IT decision-makers understand that success depends on seamless integration. This puts renewed pressure on data silos and integrations. ... For IT decision-makers wondering what they really need to do in 2026, doing nothing is definitely not the right answer, as your competitors who do invest in AI will quickly overtake you. On the other hand, you don’t have to go all-in and blow your entire IT budget on it. ... You need to start now, so start small. Putting the three or five most frequently asked questions to your customer service or HR team into an AI agent can take a huge workload off those teams. There are now several case studies showing that this has reduced the number of tickets by as much as 50-60 percent. AI can also be used for sales reports or planning, which currently takes employees many hours each week.


Mobile privacy audits are getting harder

Many privacy reviews begin with static analysis of an Android app package (APK). This can reveal permissions requested by the app and identify embedded third-party libraries such as advertising SDKs, telemetry tools, or analytics components. Requested permissions are often treated as indicators of risk because they can imply access to contacts, photos, location, camera, or device identifiers. Library detection can also show whether an app includes known trackers. Yet, static results are only partial. Permissions may never be used in runtime code paths, and libraries can be present without being invoked. Static analysis also misses cases where data is accessed indirectly or through system behavior that does not require explicit permissions. ... Apps increasingly defend against MITM using certificate pinning, which causes the app to reject traffic interception even if a root certificate is installed. Analysts may respond by patching the APK or using dynamic instrumentation to bypass the pinning logic at runtime. Both approaches can fail depending on the app’s implementation. Mopri’s design treats these obstacles as expected operating conditions. The framework includes multiple traffic capture approaches so investigators can switch methods when an app resists a specific setup. ... Raw network logs are difficult to interpret without enrichment. Mopri adds contextual information to recorded traffic in two areas: identifying who received the data, and identifying what sensitive information may have been transmitted.


When the AI goes dark: Building enterprise resilience for the age of agentic AI

Instead of merely storing data, AI accumulates intelligence. When we talk about AI “state,” we’re describing something fundamentally different from a database that can be rolled back. ... Lose this state, and you haven’t just lost data. You’ve lost the organizational intelligence that took hundreds of human days of annotation, iteration and refinement to create. You can’t simply re-enter it from memory. Worse, a corrupted AI state doesn’t announce itself the way a crashed server does. ... This challenge is compounded by the immaturity of the AI vendor landscape. Hyperscale cloud providers may advertise “four nines” of uptime (99.99% availability, which translates to roughly 52 minutes of downtime per year), but many AI providers, particularly the startups emerging rapidly in this space, cannot yet offer these enterprise-grade service guarantees. ... When AI agents handle customer interactions, manage supply chains, execute financial processes and coordinate operations, a sustained AI outage isn’t an inconvenience. It’s an existential threat. ... Humans are not just a fallback option. They are an integral component of a resilient AI-native enterprise. Motivated, trained and prepared teams can bridge gaps when AI fails, ensuring continuity of both systems and operations. When you continually reduce your workforce to appease your shareholders, will your human employees remain motivated, trained and prepared?


The blind spot every CISO must see: Loyalty

The insider who once seemed beyond reproach becomes the very vector through which sensitive data, intellectual property, or operational integrity is compromised. These are not isolated failures of vetting or technology; they are failures to recognize that loyalty is relational and conditional, not absolute. ... Organizations have long operated under the belief that loyalty, once demonstrated, becomes a durable shield against insider risk. Extended tenure is rewarded with escalating access privileges, high performers are granted broader system rights without commensurate behavioral review, and verbal affirmations of commitment are taken at face value. Yet time and again patterns repeat. What begins as mutual confidence weakens not through dramatic betrayal but through subtle realignments in personal commitment. An employee who once identified strongly with the mission may begin to feel undervalued, overlooked for advancement, or weighed down by outside pressures. ... Positions with access to crown jewels — sensitive data, financial systems, or personnel records — or executive ranks inherently require proportionately more oversight, as regulated sectors have shown. Professionals in these roles accept this as part of the terrain, with history demonstrating minimal talent loss when frameworks are transparent and supportive.


Researchers Warn: WiFi Could Become an Invisible Mass Surveillance System

Researchers at the Karlsruhe Institute of Technology (KIT) have shown that people can be recognized solely by recording WiFi communication in their surroundings, a capability they warn poses a serious threat to personal privacy. The method does not require individuals to carry any electronic devices, nor does it rely on specialized hardware. Instead, it makes use of ordinary WiFi devices already communicating with each other nearby.  ... “This technology turns every router into a potential means for surveillance,” warns Julian Todt from KASTEL. “If you regularly pass by a cafĂ© that operates a WiFi network, you could be identified there without noticing it and be recognized later, for example by public authorities or companies.” Felix Morsbach notes that intelligence agencies or cybercriminals currently have simpler ways to monitor people, such as accessing CCTV systems or video doorbells. “However, the omnipresent wireless networks might become a nearly comprehensive surveillance infrastructure with one concerning property: they are invisible and raise no suspicion.” ... Unlike attacks that rely on LIDAR sensors or earlier WiFi-based techniques that use channel state information (CSI), meaning measurements of how radio signals change when they reflect off walls, furniture, or people, this approach does not require specialized equipment. Instead, it can be carried out using a standard WiFi device.


Is software optimization a lost art?

Almost all of us have noticed apps getting larger, slower, and buggier. We've all had a Chrome window that's taking up a baffling amount of system memory, for example. While performance challenges can vary by organization, application and technical stacks, it appears the worst performance bottlenecks have migrated to the ‘last mile’ of the user experience, says Jim Mercer ... “While architectural decisions and developer skills remain critical, they’re too often compromised by the need to integrate AI and new features at an exponential pace. So, a lack of due diligence when we should know better.” ... The somewhat concerning part is that AI bloat is structurally different from traditional technical debt, she points out. Rather than accumulated cruft over time, it usually manifests as systematic over-engineering from day one. ... Software optimization has become even more important due to the recent RAM price crisis, driven by surging demand for hardware to meet AI and data center buildout. Though the price increases may be levelling out, RAM is now much more expensive than it was mere months ago. This is likely to shift practices and behavior, Brock ... Security will play a role too, particularly with the growing data sovereignty debate and concerns about bad actors, she notes. Leaner, neater, shorter software is simply easier to maintain – especially when you discover a vulnerability and are faced with working through a massive codebase.


The ‘Super Bowl’ standard: Architecting distributed systems for massive concurrency

In the world of streaming, the “Super Bowl” isn’t just a game. It is a distributed systems stress test that happens in real-time before tens of millions of people. ... It is the same nightmare that keeps e-commerce CTOs awake before Black Friday or financial systems architects up during a market crash. The fundamental problem is always the same: How do you survive when demand exceeds capacity by an order of magnitude? ... We implement load shedding based on business priority. It is better to serve 100,000 users perfectly and tell 20,000 users to “please wait” than to crash the site for all 120,000. ... In an e-commerce context, your “Inventory Service” and your “User Reviews Service” should never share the same database connection pool. If the Reviews service gets hammered by bots scraping data, it should not consume the resources needed to look up product availability. ... When a cache miss occurs, the first request goes to the database to fetch the data. The system identifies that 49,999 other people are asking for the same key. Instead of sending them to the database, it holds them in a wait state. Once the first request returns, the system populates the cache and serves all 50,000 users with that single result. This pattern is critical for “flash sale” scenarios in retail. When a million users refresh the page to see if a product is in stock, you cannot do a million database lookups. ... You cannot buy “resilience” from AWS or Azure. You cannot solve these problems just by switching to Kubernetes or adding more nodes.


Cloud-native observability enters a new phase as the market pivots from volume to value

“The secret in the industry is that … all of the existing solutions are motivated to get people to produce as much data as possible,” said Martin Mao, co-founder and chief executive officer of Chronosphere, during an interview with theCUBE. “What we’re doing differently with logs is that we actually provide the ability to see what data is useful, what data is useless and help you optimize … so you only keep and pay for the valuable data.” ... Widespread digital modernization is driving open-source adoption, which in turn demands more sophisticated observability tools, according to Nashawaty. “That urgency is why vendor innovations like Chronosphere’s Logs 2.0, which shift teams from hoarding raw telemetry to keeping only high-value signals, are resonating so strongly within the open-source community,” he said. ... Rather than treating logs as an add-on, Logs 2.0 integrates them directly into the same platform that handles metrics, traces and events. The architecture rests on three pillars. First, logs are ingested natively and correlated with other telemetry types in a shared backend and user interface. Second, usage analytics quantify which logs are actually referenced in dashboards, alerts and investigations. Third, governance recommendations guide teams toward sampling rules, log-to-metric conversion or archival strategies based on real usage patterns.


How recruitment fraud turned cloud IAM into a $2 billion attack surface

The attack chain is quickly becoming known as the identity and access management (IAM) pivot, and it represents a fundamental gap in how enterprises monitor identity-based attacks. CrowdStrike Intelligence research published on January 29 documents how adversary groups operationalized this attack chain at an industrial scale. Threat actors are cloaking the delivery of trojanized Python and npm packages through recruitment fraud, then pivoting from stolen developer credentials to full cloud IAM compromise. ... Adversaries are shifting entry vectors in real-time. Trojanized packages aren’t arriving through typosquatting as in the past — they’re hand-delivered via personal messaging channels and social platforms that corporate email gateways don’t touch. CrowdStrike documented adversaries tailoring employment-themed lures to specific industries and roles, and observed deployments of specialized malware at FinTech firms as recently as June 2025. ... AI gateways excel at validating authentication. They check whether the identity requesting access to a model endpoint or training pipeline holds the right token and has privileges for the timeframe defined by administrators and governance policies. They don’t check whether that identity is behaving consistently with its historical pattern or is randomly probing across infrastructure.


The Hidden Data Access Crisis Created by AI Agents

As enterprises adopt agents at scale, a different approach becomes necessary. Instead of having agents impersonate users, agents retain their own identity. When they need data, they request access on behalf of a user. Access decisions are made dynamically, at the moment of use, based on human entitlements, agent constraints, data governance rules, and intent (purpose). This shifts access from being identity-driven to being context-driven. Authorization becomes the primary mechanism for controlling data access, rather than a side effect of authentication. ... CDOs need to work closely with IAM, security, and platform operations teams to rethink how access decisions are made. In particular, this means separating authentication from authorization and recognizing that impersonation is no longer a sustainable model at scale. Authentication teams continue to establish trust and identity. Authorization mechanisms must take on the responsibility of deciding what data should be accessible at query time, based on the human user, the agent acting on their behalf, the data’s governance rules, and the purpose of the request. ... CDOs must treat data provisioning as an enterprise capability, not a collection of tactical exceptions. This requires working across organizational boundaries. Authentication teams continue to establish trust and identity. Security teams focus on risk and enforcement. Data teams bring policy and governance context. 

Daily Tech Digest - November 26, 2025


Quote for the day:

“There is only one thing that makes a dream impossible to achieve: the fear of failure.” -- Paulo Coelho



7 signs your cybersecurity framework needs rebuilding

The biggest mistake, Pearlson says, is failing to recognize that the current plan is out of date or simply not working. Breaches happen, but that doesn’t always mean your cyber framework needs rebuilding. It does, however, indicate that the framework needs to be rethought and redesigned. ... “If your framework hasn’t kept pace with evolving threats or business needs, it’s time for a rebuild.” Cyber threats are always evolving, so staying proactive with regular reviews and fostering a culture of cybersecurity awareness will help catch issues before they become crises, Bucher says. ... “The cybersecurity landscape has evolved rapidly, especially with the rise of generative AI — your framework should reflect these shifts.” McLeod recommends a complete a biannual framework review combined with a cursory review during the gap years. “This helps to ensure that the framework stays aligned with evolving threats, business changes, and regulatory requirements.” Ideally, security leaders should always have their security framework in mind while maintaining a rough, running list of areas that could be improved, streamlined, or clarified, McLeod suggests. ... If an organization is stuck in a cycle of continually chasing alerts and incidents, as well as reporting events after the fact instead of performing predictive threat assessments, data analysis, and forward planning, it’s time for a change, Baiati advises. 


Your Million-Dollar IIoT Strategy is Being Sabotaged by Hundred-Dollar Radios

The ambition is clear: to create hyper-efficient, data-driven operations in a market expected to exceed $1.6 billion by 2030. Yet, a fundamental paradox lies at the heart of this transformation. While we architect complex digital twins and deploy sophisticated AI models, the foundational tools entrusted to our most valuable asset—the frontline workforce—are often decades old, disconnected, and failing at an alarming rate. ... Data shows that one in four organizations loses more than an entire day of productivity every month simply dealing with broken technology. The primary culprits are as predictable as they are preventable: nearly half of workers cite battery problems (48.4%) and physical damage (46.8%) as the most common causes of failure. ... While conversations about this crisis often focus on pay and career paths, Relay’s research reveals a more immediate, tangible cause: the daily frustration of using broken tools. 1 in 4 frontline workers already feel their equipment is second-class compared to what their corporate counterparts use, and a staggering 43% of workers saying they’d be less likely to quit if guaranteed access to modern, automatically upgraded devices. ... Beyond reliability, it’s important to address the data black hole created by legacy, disconnected tools. Every day, frontline teams generate thousands of hours of spoken communication—a rich stream of unstructured data filled with maintenance alerts, safety concerns, and process bottlenecks. 


Ask the Experts: Validate, don't just migrate

"Refactoring code is certainly a big undertaking. And if you start before you have good hygiene and governance, then you're just setting yourself up for failure. Similarly, if you haven't tagged properly, you have no way to attribute it to the project, and that becomes a cost problem." ... "If you do conclude [that migration is necessary], then you really must make sure the application is architected right. A lot of times, these workloads weren't designed for the cloud world, so you must adapt them and deliberately architect them for a cloud workload. "[To prepare a mission-critical application], it's key to look at the appropriateness, operating system [and] licenses. Sometimes, there are licenses tied to CPUs or other things that might introduce issues for you as well, so regression, latency and performance testing will be mandatory. ... "[IT leaders must also understand] the risks and costs associated with taking things into the cloud, and the pros and cons of that versus leaving it alone. Because old stuff, whether it was [procured] yesterday or five years ago, is inherently going to be vulnerable from a cybersecurity standpoint. Risk No. 2 is interoperability and compatibility, because old stuff doesn't talk to new stuff. And the third one is supportability, because it's hard to find old people to support old systems. ... "Sometimes, people have the false sense that if it's in cloud, then I'm all set. Everything is available, and everything is highly redundant. And it is, if you design [the application] with those things in mind.


Heineken CISO champions a new risk mindset to unlock innovation

Starting as an auditor and later leading a cyber defense team. It’s easy to fall into the black-and-white trap of being the function that always says “no” or speaks in cryptic tech jargon. It’s a scary world out there with so many attacks happening in every industry. The classical reaction of most security professionals is to tighten defences and impose even more rules. ... CISOs need to shift the mindset from pure compliance to asking: How does our cyber strategy support the business and its values? What calculated risks do we want the business to take? Where do we need their attention and help to embed security into the DNA of our people and our company? ... Be visible and approachable. Share the lessons that shaped you as a leader, what worked, what didn’t, and the principles that guide your decisions. I’m passionate about building diverse teams where everyone gets the same opportunities, no matter age, gender, or background. Diversity makes us stronger, and when there’s trust and openness, it sparks mentoring, coaching, and knowledge sharing. Make coaching and mentoring non-negotiable, and carve out time for it. It’s easy to push aside when you’re busy putting out security fires, but neglecting people’s growth and well-being is a big miss. Be authentic and vulnerable, walk the talk. Share the real stories, including failures and what made you stronger. Too often, people focus only on titles, certifications, and tech skills.


Data-Driven Enterprise: How Companies Turn Data into Strategic Advantage

A data-driven enterprise is not defined by the number of dashboards or analytics tools it owns. It’s defined by its ability to turn raw information into intelligent action. True data-driven organizations embed data thinking into every level of decision-making from boardroom strategy to day-to-day operations. ... A modern data architecture is not a single platform, but an interconnected ecosystem designed to balance agility, governance, and scalability. ... As organizations mature in their data journey, they are moving away from rigid, centralized models that rely on a single source of truth. While centralization once ensured control, it often created bottlenecks slowing down innovation and limiting agility.  ... We are entering an era of data agents self-learning systems capable of autonomously detecting anomalies, assessing risks, and forecasting trends in real time. These intelligent agents will soon become the invisible workforce of the enterprise, operating across domains: predicting supply chain disruptions, optimizing IT performance, personalizing customer journeys, and ensuring compliance through continuous monitoring. Their actions will reshape not only operations but also how organizations think about governance, accountability, and human oversight. For architects, this shift represents both a challenge and an extraordinary opportunity. The role is evolving from that of a data custodian focused on structure and governance to an ecosystem designer who engineers environments where data and AI can coexist, learn, and continuously create value.


10 benefits of an optimized third-party IT services portfolio

By entrusting day-to-day IT operations to trusted providers, organizations can reallocate internal resources toward higher-value initiatives such as digital transformation, automation, and product innovation. This accelerates adoption of emerging technologies, and allows internal teams to deepen business expertise, strengthen cross-functional collaboration, and focus on driving growth where it matters most. ... A well-structured third-party IT services portfolio can provide flexibility to scale up or down based on business needs. This is particularly valuable for CEOs who need to adapt to changing market conditions and seize growth opportunities. Securing talent in the market today is challenging and time consuming, so tapping into the talent pools of your strategic IT services partner base allows organizations to leverage their bench strength to fill immediate needs for talent. ... IT service providers continuously invest in advanced tech and talent development, enabling clients to benefit from cutting-edge innovations without bearing the full cost of adoption. As AI, automation, and cybersecurity evolve, providers offer the subject matter expertise and tools organizations need to stay ahead of disruption. ... With operational stability ensured through a balance of internal talent and trusted third parties, CIOs can dedicate more focus to long-term strategic initiatives that fuel growth and innovation. 


Modernizing SOCs with Agentic AI and Human-in-the-Loop: A Guide to CISOs

Traditional SOCs were not built for today’s speed and scale. Alert fatigue, manual investigations, disconnected tools, and talent shortages all contribute to the operational drag. Many security leaders are stuck in a reactive loop with no clear path to improvement. ... Legacy SOCs rely heavily on outdated technologies and rule-based detection, generating high volumes of alerts, many of which are false positives, leading to analyst burnout. Analysts are compelled to manually inspect and triage a deluge of meaningless signals, making the entire effort unsustainable. ... Before transformation can happen, one needs to understand where one stands. This can be accomplished with key benchmarking metrics for SOC performance, such as MTTD (Mean time to detect), MTTR (Mean time to respond), case closure rates, and tool effectiveness. ... Agentic AI represents the next evolution of AI-powered cybersecurity, which is modular, explainable, and autonomous. Through a coordinated system of AI agents, the Agentic SOC continuously responds and adapts to the evolving security environment in real time. It is designed to accelerate threat detection, investigation, and response by 10x, bringing speed, precision, and clarity to every function of SecOps. Agentic AI is the technology shift that changes the game. Unlike traditional automation, Agentic AI is decision-oriented, self-improving, and always operating with human-in-the-loop for oversight.


3 SOC Challenges You Need to Solve Before 2026

2026 will mark a pivotal shift in cybersecurity. Threat actors are moving from experimenting with AI to making it their primary weapon, using it to scale attacks, automate reconnaissance, and craft hyper-realistic social engineering campaigns. ... Attackers have mastered evasion. ClickFix campaigns trick employees into pasting malicious PowerShell commands by themselves. LOLBins are abused to hide malicious behavior. Multi-stage phishing hides behind QR codes, CAPTCHAs, rewritten URLs, and fake installers. Traditional sandboxes stall because they can't click "Next," solve challenges, or follow human-dependent flows. Result? Low detection rates for the exact threats exploding in 2025 and beyond. ... Thousands of daily alerts, mostly false positives. An average SOC handles 11,000 alerts daily, with only 19% worth investigating, according to the 2024 SANS SOC Survey. Tier 1 analysts drown in noise, escalating everything because they lack context. Every alert becomes a research project. Every investigation starts from zero. Burnout hits hard. Turnover doubles, morale tanks, and real threats hide in the backlog. By 2026, AI-orchestrated attacks will flood systems even faster, turning alert fatigue into a full-blown crisis. ... From a financial leadership perspective, security spending often feels like a black hole: money is spent, but risk reduction is hard to quantify. SOCs are challenged to justify investments, especially when security teams seem to be a cost center without clear profit or business-driving impact.


Digital surveillance tools are reshaping workplace privacy, GAO warns

Privacy concerns intensify when surveillance data feeds into automated systems that evaluate performance, set productivity metrics, or flag workers for potential discipline. GAO found that employers often rely on flawed benchmarks and incomplete measurements. Tools rarely capture the full range of work performed, such as research, mentoring, reading, or off-screen tasks, and frequently misinterpret normal behavior as inefficiency. When employers trust these tools “at face value,” the report notes, workers can be unfairly labeled unproductive or noncompliant despite doing their jobs well. ... Meanwhile, past federal efforts to issue guidance on reducing surveillance related harms such as transparency practices, human oversight, and safeguards against discriminatory impacts have been rescinded or paused since January by the Trump administration as agencies reassess their policy priorities. GAO also notes that existing federal privacy protections are narrow. The Electronic Communications Privacy Act restricts covert interception of communications, but it does not cover most forms of digital monitoring, such as keystroke logging, location tracking, biometric data collection, or algorithmic productivity scoring. ... The report concludes that while digital surveillance can improve safety, efficiency, and health monitoring, its benefits depend wholly on how employers use it.


How to avoid becoming an “AI-first” company with zero real AI usage

A competitor declared they’re going AI-first. Another publishes a case study about replacing support with LLMs. And a third shares a graph showing productivity gains. Within days, boardrooms everywhere start echoing the same message: “We should be doing this. Everyone else already is, and we can’t fall behind.” So the work begins. Then come the task forces, the town halls, the strategy docs and the targets. Teams are asked to contribute initiatives. But if you’ve been through this before, you know there’s often a difference between what companies announce and what they actually do. Because press releases don’t mention the pilots that stall, or the teams that quietly revert to the old way, or even the tools that get used once and abandoned. ... By then, your company’s AI-first mandate will have set into motion departmental initiatives, vendor contracts and maybe even some new hires with “AI” in their titles. The dashboards will be green, and the board deck will have a whole slide on AI. But in the quiet spaces where your actual work happens, what will have meaningfully changed? Maybe you'll be like the teams that never stopped their quiet experiments. ... That’s invisible architecture of genuine progress: Patient, and completely uninterested in performance. It doesn't make for great LinkedIn posts, and it resists grand narratives. But it transforms companies in ways that truly last. Every organization is standing at the same crossroads right now: Look like you’re innovating, or create a culture that fosters real innovation.

Daily Tech Digest - August 24, 2023

3 data privacy principles to adopt now, even while governments still debate

Fairness is one of the most powerful guiding principles any brand can adopt for its use of data, but what does it mean in practice? On the one hand, it’s about considering how you’re using not just data but the tools and technologies that help you harness data in your marketing and decision-making. On the other hand, it’s important to remember we’re not just talking about one moment in time, like the moment when someone gives you their data, or the moment of an interaction between them and you, in a store or on your website. It’s about the potential implications that these moments can have down the line. Could it lead to an unfair, harmful, or discriminatory outcome for them? Could it keep them from getting credit? Or a job offer? Could it perpetuate a stereotype about a protected class of people? Building a foundation of fairness, for example, could mean implementing policies and procedures to regularly assess the data and tech you use to ensure they do not have a disparate impact on vulnerable consumers.


Cyber attackers using Gen AI more effectively than defenders

Both cyber attackers and defenders employ generative AI, but attackers use it more effectively. Adversaries capitalise on AI/ML, deepfake, facial recognition, and Augmented Reality/Virtual Reality (VR) (AR/VR) to enhance hacking strategies against government agencies, businesses, and strategic targets, surpassing cyber defenders in technological adaptation. Facial recognition and AR/VR systems illustrate the extensive use of deepfake technology by cybercriminals. We predict that within two years, social engineering and phishing attacks will predominantly employ deep fakes, making defenders' tasks much harder. Malware capabilities have evolved significantly. Instead of creating static malware, hackers now build multi-behavioural malware that adapts in real-time. Upon reaching a target, this malware assesses the environment and generates tailored malicious code, targeting various systems like Windows, Linux, Outlook, and mobile devices. This is powered by AI/ML engines, resulting in multi-behavioural, metamorphic, and polymorphic malware that dynamically alters their code as they spread.


Cloud Robotics: A New Frontier for Internet Technology

Robots connected to the cloud are being used in warehouses and distribution centers for material handling, order fulfillment, and inventory management duties. These robots are capable of independent navigation, object recognition and picking, and teamwork with human personnel. The medical sector is likewise ripe for transformation because to cloud robots. Robots connected to the cloud can access patient information, medical records, and cutting-edge disease-diagnosis algorithms. Cloud robotics alters how we connect with our domestic environment regarding home automation. Robots with cloud capabilities can automate harvesting, monitor crop health, and manage resource usage in agriculture. These robots can use the cloud to evaluate massive volumes of field data, forecast agricultural yields, and make quick judgments. Cloud robotics has tremendous promise as we look to the future. Advanced artificial intelligence (AI) and cloud robotics are being combined as a new trend, allowing robots to act more intelligently and quickly adapt to their surroundings.


Organizing Around Business Capabilities

A Value Structure is an idealized teaming structure illustrating how the organization delivers benefits to its customers. The idealized structure includes teams and roles to not only operate a capability, but also to build it. We call this structure the value structure to differentiate it from two other structures within an organization: formal structure and learning structure. The formal structure represents the way an organization structures its activities into jobs and job families, manages compensation and other aspects of human resources. The learning structure represents the way an organization learns to improve its performance, including role-based learning, team-based learning, and establishing a culture of relentless improvement without guilt or blame. Establishment of a value structure independent from formal and learning structures enables an organization to begin to change how it delivers value to customers without the overhead of changing formal reporting or job titles. The value structure makes impediments to the flow of value clearly visible so we can either eliminate them or explicitly orchestrate them.


How to Build True Cyber Resilience

Cyber resilience cannot be achieved by implementing one initiative or investing in one new technology. “CISOs should focus on the question, ‘How ready are we?’" says Hopkins. Are organizations ready to detect threats, respond to them, recover, and adapt to an ever-changing threat landscape? “The first step to building cyber resilience involves understanding which cyberattacks are most relevant to an organization based on its industry, location, IT ecosystem, data type, users, etc.,” says Tony Velleca, CISO at digital technology and IT service company UST and CEO of CyberProof, a UST security services company. Once an organization understands its risks, the question becomes how to detect those threats, stop them, and contain them if and when they become cybersecurity incidents. The answer lies in a blend of technology and talent. Combining the power of cybersecurity tools, such as zero trust and managed detection and response, can help organizations achieve cyber resilience, but they need to ensure the strategies they deploy make measurable progress toward that goal.


AI and the evolution of surveillance systems

AI models are influenced by the datasets used to train them. It is imperative that AI vendors carefully tune and balance their datasets to prevent biases from occurring. Balancing datasets is a manual process that requires making sure that the humans visible in the datasets are a good representation of reality, and do not have biases towards certain human traits. In our case, we use diverse groups of actors, from all over the world, to play out violence for our training datasets to ensure they are balanced. Furthermore, testing regularly for such biases can go a long way. A carefully designed system can protect and help people without significantly impacting their privacy. This requires considering privacy from designing to implementing AI systems. I believe that the future of AI-powered surveillance will see reduced privacy infringement. Currently, large surveillance installations still require humans looking at camera streams all the time. In a trigger-based workflow, where humans take actions after an AI has alerted them, the amount of security camera footage seen by humans is much less, and thus the risk of privacy infringement decreases.


Controversial Cybercrime Law Passes in Jordan

A joint statement by Human Rights Watch, Access Now, Article 19, and 11 other organizations said the bill has several provisions threatening freedom of expression, the right to information, and the right to privacy, as well as tightening government control over the Internet. The groups also claimed the bill will introduce new controls over social media, weaken online anonymity, hamper free expression and access to information, and increase online censorship. Meantime the European Union says it recognizes and supports Jordan's objective to create a strong legislative framework to deal with and counter cybercrime efficiently, but it contends that some of the provisions of the new cybercrime law depart from international human rights standards and could result in limiting freedom of expression online and offline. Liz Throssell, the United Nations' spokesperson for the UN High Commissioner for Human Rights, said countries indeed need to take steps to combat cybercrime, but protecting security online and ensuring online freedoms must be treated as complementary goals.


Evaluating Open Source: Green Flags to Look For

First and foremost, is the open-source community for the solution vibrant; is it widely adopted and does the community regularly contribute updates? A healthily engaged community is a sign that the technology has legs and that companies are successful with it; it often indicates the extent to which companies are employing staff to contribute to the community. Closely related to this point, does the open source technology actually solve the problems you need solved? With the enormous popularity of open source comes the enormous hype around novel technologies, but are those technologies actually something that help solve your business problems in a sustainable way such that you can be confident that your investments may carry you several years? You should evaluate the suitability of open source technology in the same way you evaluate proprietary technology and not let the free or low-cost factors lead to hasty decisions. Finally, are vendors providing software, services, and support for the open source technology? 


How Threat Research Can Inform Your Cloud Security Strategy

The most important thing to remember about cybersecurity is that it’s not an action you take, but a practice you follow. Implementing a strong cloud security posture requires regularly assessing and updating your cloud security policies in light of new threats or not. This means being proactive in your protection strategies and planning for the unexpected. Creating an incident response plan is a great place to start, and continuing employee education and training will help embed a security-focused mindset across the organization as a whole. There is no “one right way” to establish a cloud security strategy, but it’s a sure bet that being informed is a good move. Keeping up to date on the latest cybersecurity threats and vulnerabilities through sources like the National Vulnerability Database and Orca Research Pod is a good place to start. However, proactive measures like implementing best practices, organizational training, and even bug bounties and other security policies can go a long way toward creating a well-informed cloud security posture.


Regulatory uncertainty overshadows gen AI despite pace of adoption

In traditional application development, enterprises have to be careful that end users aren’t allowed access to data they don’t have permission to see. For example, in an HR application, an employee might be allowed to see their own salary information and benefits, but not that of other employees. If such a tool is augmented or replaced by an HR chatbot powered by gen AI, then it will need to have access to the employee database so it can answer user questions. But how can a company be sure the AI doesn’t tell everything it knows to anyone who asks? This is particularly important for customer-facing chatbots that might have to answer questions about customers’ financial transactions or medical records. Protecting access to sensitive data is just one part of the data governance picture. “You need to know where the data’s coming from, how it’s transformed, and what the outputs are,” says Nick Amabile, CEO at DAS42, a data consulting firm. “Companies in general are still having problems with data governance.”



Quote for the day:

"The leader has to be practical and a realist, yet must talk the language of the visionary and the idealist." -- Eric Hoffer

Daily Tech Digest - October 03, 2022

Roadmap to RPA Implementation: Thinking Long Term

Ted Kummert, executive vice president of products and engineering at UiPath, says RPA should be viewed as a long-range capability meant to empower organizations to evolve strategically and increase business value. It is a journey that can start small, within one division or one department, and grow organically across the business as additional ideas form and the organization’s vision for automation’s potential comes to fruition. He says RPA can clear backlog, create new capacity, and free up resources, and improve data quality by integrating software robots into workflows. “It is a truly transformative technology that can reduce or eliminate manual tasks and elevate creative, high-value work,” Kummert says. “Digital transformation is often talked about, but many times can fall short of its goals. Automation is the driver to achieve true digital transformation.” Adam Glaser, senior vice president of engineering for Appian, says many businesses use one automation technology, adding third-party capabilities in patchwork fashion to automate complex end-to-end processes.


How to start and grow a cybersecurity consultancy

To be successful, an entrepreneur must be resilient. Any comment that runs along the lines of “That’s not possible,” or “That can’t be done” should be treated as a challenge to prove the speaker wrong. An entrepreneur needs to have the ability to see through what’s not important. Entrepreneurs don’t just need money – they also need support in the form of encouragement and advice. I would advise budding entrepreneurs to attend meetups within their industry or local community and seek out online support via forums and groups. You’ll be surprised just how willing others will be to help and offer advice for free. Asking questions, getting reassurance and sanity checks from peers can be invaluable at all stages of your businesses journey. There will always be someone a little further down the path you’re taking. Starting a business can be exhilarating, rewarding and fun, but can be exhausting, relentless and stressful in equal measure.


Surveillance tech firms complicit in MENA human rights abuses

“When operating in conflict-affected or high-risk regions as the MENA region, the surveillance sector must undertake heightened human rights due diligence and, if it cannot do so or it identifies evidence of harm, it should stop selling its technology to companies or governments,” said Dima Samaro, MENA regional researcher and representative at the Business & Human Rights Resource Centre. “Lack of adequate due diligence measures by private companies will only worsen the situation for those from marginalised communities, putting their lives in jeopardy as the absence of robust regulation and effective mechanisms in the region allows surveillance technologies to be operated freely and without scrutiny.” The report added that, although the United Nations’ (UN) Guiding principles on business and human rights were adopted a decade ago – which establish that companies must take proactive and ongoing steps to identify and respond to the potential or actual human rights impacts of their business – the principles’ non-binding, voluntary nature means there are “glaring gaps in human rights safeguards” at the firms.


How companies can accelerate transformation

Ensuring that customer value drives technology architecture and investment is one way to optimize technology usage. Another way is to ensure that an organization is getting the most out of the investments it has already made. Inefficiency in any aspect of technology usage represents a drag on businesses’ ability to change quickly. ... While enterprise architects (EAs) play a central role in identifying opportunities for this type of technology optimization, they have an even greater role to play when it comes to optimizing the entire IT landscape. A “business capability” perspective makes this possible. ... Efficiency doesn’t improve on its own. The business needs to decide to improve it. Making those decisions, however, is not always easy. As mentioned, relying on business capabilities to evaluate technology needs is one way to simplify the decision process. The other is visibility. Business leaders can’t make decisions if they can’t see the problem. In terms of business architecture, EAs help guide leaders in the decisions they make by showing them business capability maps, data-rich process diagrams and dashboards highlighting the connection between architectural issues and business value.


Optus reveals extent of data breach, but stays mum on how it happened

Optus says its recent data breach impacted 1.2 million customers with at least one form of identification number that is valid and current. The Australian mobile operator also has brought in Deloitte to lead an investigation on the cybersecurity incident, including how it occurred and how it could have been prevented. Optus said in a statement Monday that Deloitte's "independent external review" of the breach would encompass the telco's security systems, controls, and processes. It added that the move was supported by the board of its parent company Singtel, which had been "closely monitoring" the situation. Elaborating on Deloitte's forensic assessment, Optus CEO Kelly Bayer Rosmarin said: "This review will help ensure we understand how it occurred and how we can prevent it from occurring again. It will help inform the response to the incident for Optus. This may also help others in the private and public sector where sensitive data is held and risk of cyberattack exists." In its statement, Optus added that it had worked with more than 20 government agencies to determine the extent of the data breach.


Why cyber security strategy must be more than a regulatory tick-box exercise

While technology plays a critical role in an effective cyber security strategy, it alone does not provide the solution. Business leaders must also consider the organisation’s processes and people. If organisations don’t have the right processes or people in place to manage new technologies, it can be easy to revert to old habits. Many organisations opt for a hybrid Security Operations Centre to underpin their MDR strategy, which combines the cyber skills of in-house engineers, cyber security teams and an MSSP to create a single facility. MSSPs fill in the gaps in defences while upskilling in-house teams to stay on top of changing threats and technologies. This approach can also free in-house staff to drive projects and internal improvements while the MSSP takes the lead on high value incidents. If the goal is to improve cyber security whilst meeting your organisational goals, then regulations will only ever go so far in tackling the issue. Attacks will continue to plague all sectors and proper detection, response and remediation will be what makes the difference between those that make the news and those that don’t.


Mozilla is looking for a scapegoat

Not so long ago, Microsoft’s Internet Explorer dominated market share. Antitrust authorities helped change that, but Google, not Mozilla, stepped up to take Microsoft’s place, yet without the bully pulpit of a dominant operating system. Meanwhile, as far back as 2008, I was writing about Mozilla’s chance to make Firefox a true community-developed web platform. It didn’t succeed, though Mozilla has gifted us incredible innovations such as Rust. Clearly there are smart people at Mozilla and they have demonstrated the ability to push the envelope on innovation. But not with Firefox. DuckDuckGo has carved out a growing, sizeable niche in privacy-oriented search, but Mozilla keeps losing similar ground in browsers. Why? In its report, Mozilla says browser freedom has been “suppressed for years through online choice architecture and commercial practices that benefit platforms and are not in the best interest of consumers, developers, or the open web.” This would be more credible in Mozilla’s mouth if this weren’t the same company that completely mismanaged its entrance into the mobile market.


Indonesia Data Protection Law Includes Potential Prison Time

The Indonesia data protection law took some eight years to come to fruition, with contentious ongoing debate about what government body should oversee the new regulations and exactly how strong the penalties should be. A recent wave of cyber attacks and data breaches in the country seems to have prompted legislative action; Kaspersky reports that the country experienced 11.8 million cyberattacks in the first quarter of 2022, a 22% increase from the prior year, and the country has become the leading target for ransomware attacks in Southeast Asia. This includes data breaches of various government agencies, one of which exposed the vaccination records of President Joko Widodo. Stats from SurfShark indicate that Indonesia now has the third-highest rate of data breaches in the world. Regulation oversight has fallen to the executive branch, with the President slated to form an oversight body tasked with determining and administering fines. Similar to the EU’s General Data Protection Regulation (GDPR), which the Indonesia data protection law drew from substantially, there is a maximum potential fine of 2% of global annual turnover for violations.


How To Protect Your Reputation After A Hack Or Data Breach

Part of transparency and recovery is working with the relevant authorities and experts to track the scope of the breach. A post-mortem analysis can be critical. For one thing, it can determine what data was stolen, by who and how. It can also help track where that data ends up and how it is used. In cases where the cause has something to do with software or hardware being exploited, it can be essential to inform the developers or manufacturers of the breach and how it occurred. They may also need to issue patches or recalls to prevent other businesses using that hardware or software from being compromised. No business stands alone. ... Recovery after a breach is a sensitive time. You will undoubtedly see a deluge of negative reviews and bad press, which will be difficult to counteract. Clear and transparent messaging is part of it; breaches happen, and there's no surefire way to avoid them. Demonstrating that your data security policies prevented usable data from being stolen or that you've been able to protect users proactively can be critical to repairing your reputation.


Data quality is at the heart of successful data governance

The downstream effects of data quality have ramifications felt throughout data governance efforts. Recent findings from a survey by Enterprise Strategy Group showed that data management is greatly challenged by a lack of visibility and compounded by data quality issues. Concerningly, 42 percent of all respondents indicated at least half of their data was “dark data” - retained by the organization, but unused, unmanageable, and unfindable. An influx in dark data and a lack of data visibility often leads to downstream bottlenecks, impeding the accuracy and effectiveness of operational data. Data quality was the top driver for organizations’ data governance programs but was also the top challenge that these organizations have to overcome to maximize the return on their data governance efforts. When you consider the fact that many organizations are experiencing data quality issues, which are difficult to manage, and in many cases have significant amounts of data that is dark, there is a clear need for more robust data governance solutions providing data landscape transparency united with business context and guidance.



Quote for the day:

"Perhaps the ultimate test of a leader is not what you are able to do in the here and now - but instead what continues to grow long after you're gone" -- Tom Rath

Daily Tech Digest - June 21, 2020

Core systems strategy for banks

There are two main options (with a few variations) for banks that conclude that they need to replace their core banking system: a traditional enterprise core banking system (self-hosted or as a utility) and a next-generation cloud-based core banking system. Most current implementations are still of the traditional variety. But we are seeing an increase in banks of all sizes putting off traditional core implementations with the aim of experimenting with next-gen systems. There is some evidence to suggest that banks will try and shift en masse to a cloud-based microservice architecture in the next few years. The core method of communication between machines will be APIs. Armed with a micro-service based architecture, the new core banking applications will become core enablers of the shift to this architecture. Traditional core banking providers have become aware of the need and potential inherent in a cloud-based microservice architecture; banking leaders should keep a close watch on developments here. We also expect to see some M&A activity between traditional and next-gen core banking system providers.


Cybersecurity In The M&A Process: A CISO's Strategy

IT departments and information security professionals are traditionally not included in the discussions leading into a merger or acquisition and are usually not given the liberty to conduct their own assessments prior to M&A execution. This can lead to a dramatic increase in cyber risks or, even worse, inheriting compromised networks. With the rapid scaling of organizations in the world of M&A, it can become exponentially more difficult to control cybersecurity risks when information security departments are already struggling to keep attackers at bay with the limited personnel and resources they have. However, there are strategies that can help get information security professionals into business conversations regarding M&As. If the cards are played correctly, this can lead to positive financial and cybersecurity outcomes. Develop a proactive plan within your organization to leverage cybersecurity as a tool at the negotiation table for the M&A process. The equation is simple: If your organization inherits a compromised network or an organization that has a poor security posture, this will cost you extra dollars that are unseen through the lens of traditional M&A cost calculations.


North Korean state hackers reportedly planning COVID-19 phishing campaign targeting 5M across six nations

SingCERT confirmed it received "information regarding a potential phishing campaign" and, in response, posted an advisory on its website Friday. It said there were "always" ongoing phishing attempts by various cybercriminals that used different themes and baits and spoofed different entities. This tactic remained a common and effective technique used to gain access to individuals' accounts, deliver malware, or trick victims into revealing confidential data, said SingCERT, which sits under Cyber Security Agency (CSA). ZDNet asked the government agency several questions including whether there had been a database breach and what tools the Manpower Ministry had adopted to prevent their email accounts from spoofing attacks. It did not respond specifically to any of the questions and, instead, issued a response that confirmed CSA had reached out to relevant parties to notify them about the potential phishing campaign. "Opportunistic cybercriminals have been using the COVID-19 situation to conduct malicious cyber activities and with the increasing reliance on the internet during this period, it is important to be vigilant," the agency said


CIA Finds It Failed to Secure Its Own Systems

The report calls out the CIA's Center for Cyber Intelligence for not prioritizing internal cybersecurity and focusing, instead, on developing offensive cyber weapons. This lax attitude toward preventive cybersecurity measures within the CIA continued even after previous high-profile data breaches of the agency and other intelligence departments, the report states. On Tuesday, Wyden wrote to John Ratcliffe, the director of national intelligence, demanding to know if the U.S. intelligence community planned to implement better cybersecurity practices and questioning why the CIA did not do more to protect its internal security operations from both outside attacks and internal threats. "The lax cybersecurity practices documented in the CIA's WikiLeaks Task Force report do not appear to be limited to just one part of the intelligence community," Wyden writes. "The Office of the Inspector General of the Intelligence Community revealed in a public summary of a report it published last year that it found a number of deficiencies in the intelligence community's cybersecurity practices."


Cyber Security Careers Germany – Finding New Roles in a Burgeoning Sector

From machine learning to autonomous response, cyber security is a burgeoning space and this is creating opportunities across Germany, from Berlin and Frankfurt to Cologne, Munich and Hamburg. Whether local markets are largely comprised of businesses still in lockdown or those that have returned to socially distanced office environments, Glocomms Germany expert consultants are able to ensure that organisations are able to meet their recruitment needs and individuals can begin planning career-defining moves. As the business world continues to adapt to the impact of COVID-19 on networks and systems, cyber security remains at the top of the agenda across sectors. Luis Rolim, Chief Marketing Officer at Glocomms commented "As the world emerges from the COVID-19 pandemic, Glocomms remains at the forefront of delivering quality talent to the technology sector. We're in this together and we look forward to helping businesses across Germany with their recruitment and talent acquisition." Glocomms Germany is part of the Phaidon International group and is a trusted recruitment partner in Europe and beyond.


What is emotion AI and why should you care?

One of the areas of emotion AI is sentiment analysis, a field that has existed since at least the early 2000s. Sentiment analysis is usually conducted on textual data, be it emails, chats, social media posts, or survey responses. It uses NLP, computational linguistics, and text analytics to infer positive or negative attitudes (aka “orientation”) of the text writer: Do they say good or bad things about your brand and your products or services? The obvious applications of sentiment analysis have been brand/reputation management (especially on social media), recommender systems, content-based filtering, semantic search, and understating user/consumer opinions, and the need to inform product design, triaging customer complaints, etc. Several of the conference presentations were devoted to this topic, which, despite all the recent progress in NLP and related fields, is still hard. Not least because there is little agreement among researchers on even what constitutes basic human emotions and how many of them are there, said Bing Liu, Professor of Computer Science at the University of Illinois at Chicago. Emotions are also notoriously hard to identify and code (label), since they are ambiguous, shifting, overlapping, and adjacent. For example, one can feel anger, sadness, and disgust at the same time. Moreover, emotions are not always easy to pin down.


Security surprise: Four zero-days spotted in attacks on researchers' fake networks

To examine the security threats to industrial systems, the researchers used a network of 120 high-interaction honeypots – fake industrial infrastructure – in 22 countries to mimic programmable logic controllers and remote terminal units. Over a period of 13 months, there were 80,000 interactions with the honeypots – mostly scans – and nine interactions that made malicious use of an industrial protocol. While that might sound like a small number, four of the nine interactions also featured previously unknown attacks, or zero-days, one being the first use of a previously identified proof-of-concept attack in the wild. The attack types include denial-of-service and command-replay attacks. These vulnerabilities and associated exploits were disclosed to the device manufacturers. "While the yield was small, the impact was high, as these were skilled, targeted exploits previously unknown to the ICS community," the researchers said. The research was presented at a NATO-backed cybersecurity conference.


Revised DOJ compliance guidance offers risk-management lessons for cybersecurity leaders

“One of the reasons the DOJ puts this out is to help compliance officers and security teams and people who are worried about bribery and corruption to ensure that the board and leadership give enough attention to these issues and properly fund them to mitigate risk,” Penman says. Regardless of whether civil or criminal litigation is involved, the kind of guidance DOJ puts out is devoured by compliance officers across all organizations, Penman says, and when it comes to compliance, cybersecurity is top of mind for those executives. “We’re just about to publish results of the survey of around 1,400 compliance officers. The highest priority or concern for risk compliance programs in that survey was enhancing data privacy and cybersecurity and data protection.” Compliance programs are more critical than ever given the COVID-19 crisis, Alison Furneaux, vice president of marketing for cybersecurity compliance management company CyberSaint, tells CSO. “The attack surface has expanded dramatically. Organizations are being forced to innovate. They’re being forced to put into place processes that they didn’t have before. They’re being forced to document and prepare for audits in a much more proficient way.”


The Difference Between Enterprise Architecture and Solutions Architecture

Perhaps it’s misleading to use “versus” to describe the difference between enterprise architecture and solutions architecture. They are very much collaborators in the organization and should not be looked at as competitive in terms of which provides more value. A better way of highlighting the difference between the two is through their focus on strategy vs. technology. A focus on strategy implies a broad understanding of the mechanics of any given technology. This is because there is a lot more to strategy than just the technology needed to implement it. A skewed focus on technology would mean that the processes, people and other variables required to inform strategy are ignored. Conversely, a focus on technology is necessary to ensure implementations and operations can run smoothly. By its nature, it is more “in the weeds” and so the necessary holistic perspective of the organization can be harder to understand and/or account for. With their holistic view of the organization, enterprise architects take on the strategy. They then use their strategic planning perspective to inform and delegate to solutions architects.


Police ties to Ring home surveillance come under scrutiny

The idea of cameras in police investigations isn’t new. Grainy black-and-white footage has been used for surveillance for years. But newer products that cost as little as $100 and connect with a cellphone make the market much more accessible. And the more people have the cameras, the more appealing their potential becomes for police and government officials. More localities are joining the registry trend. At least 75 police departments and municipalities in 21 states announced programs since 2018, according to a Stateline review. “I do think for law enforcement it’s easy to understand the appeal,” said Lior Strahilevitz, a professor at the University of Chicago’s Law School. “There are a lot of instances where if only there had been a bystander on that corner at that time, the crime could have been solved.” The registries come in a variety of forms — some a simple spreadsheet, others a more sophisticated account with vendors such as a Motorola-run program called CityProtect. (A Motorola spokeswoman declined to give a specific number but said “hundreds” of police agencies use its CityProtect service for registering cameras and/or reporting crime.) The registries can include any kind of camera from Ring to Nest to lesser known brands.



Quote for the day:

"The highest reward for a man's toil is not what he gets for it but what he becomes by it." -- John Rushkin

June 09, 2015

Are you prepared for the future of data centers?
Colocation requires a shift in data center skillsets, Koppy noted, not handing the data center over to a third party. Ask questions -- specifics about the colocation provider's network and power paths and so on -- and if the colocation provider is unwilling to share information your own facilities team would know, consider that a red flag, Courtemanche said. Also, talk to the provider's long-term customers to gauge how your own experience might be. ... There are two problem areas data centers with more than 1,000 servers experience at a much higher rate than smaller ones, according to survey results from IDC: downtime due to human error and security breaches. As one AFCOM Symposium attendee put it, when you outsource, your job goes from managing the data center to managing the colocation provider.


The top 10 myths about agile development
To be flexible has become vital for a business in today’s global markets, and therefore, the ability for IT systems to be equally flexible is essential. The purpose of agile is to allow organisations to react to the increasingly dynamic opportunities and challenges of today’s business world, in which IT has become one of the key enablers. Agile is defined by four values and 12 principles found in the Agile Manifesto. The manifesto provides an umbrella definition, in which there are many other delivery and governance frameworks, such as Scrum or extreme programming, for example.


Is Nepotism Undermining Your Business Technology Innovation?
We no longer do the break-fix relationship. We have a strategy manager that essentially acts as a CIO and manages technology as our clients grow and innovate. You need someone to be there every time you grow and change out a piece of technology and that person needs to have extensive experience throughout your industry with companies of all sizes. A small company that is a family friend doesn’t have that kind of expertise. ... Most “family friend” businesses don’t have this in place and have no idea what sort of support their users are getting, how the response time is or which issues are being resolved and escalated. You don’t have the capital to pay your users to hang out waiting for a call back on an issue.


Erasure Coding For Fun and Profit
Erasure coding essentially uses maths to add a little bit of extra data to the end of the actual data so that if you lose part of this new, bigger amount of data, you can still get all of the original data back. A simple version is a checksum: sum all the ones and zeros and put that at the end. If you lose any one of the bits, you can figure out what it was by re-calculating the checksum and comparing it to the stored checksum. The difference is what the bit was, basically. This is a vast over-simplification, but that’s basically it. ...  There’s a downside (there’s always a downside). If you lose a disk, you have to rebuild all the data from the parity blocks scattered around the place, which reduces the performance of the array because some of the time is spent on the rebuild instead of serving up the data.


Obama vows to boost U.S. cyber defenses amid signs of China hacking
"We have to be as nimble, as aggressive and as well-resourced as those who are trying to break into these systems," Obama told a news conference at the Group of Seven (G7) summit in Germany. U.S. officials, speaking on condition of anonymity, have blamed Chinese hackers for breaching the computers of the Office of Personnel Management and compromising the records of up to four million current and former employees in one of the biggest known attacks on U.S. federal networks. The mission of the intruders, the officials said, appears to have been to steal personal information for recruiting spies and ultimately to seek access to weapons plans and industrial secrets.


Rise of the Surveillance Platform
Hildyard likened a trade-surveillance platform to a buy-and-build hybrid. Such a system requires customization to effectively detect and prevent abuse, as each market ecosystem is unique. But at the same time, building the capability from the ground up is unrealistic. Delivering surveillance via a platform rather than an application gives developers leeway to develop code that’s unique to their organization and the types of behaviors they need to monitor. Sell-side banks “can’t rely on an application to do that,” Hildyard said. “The frequency with which regulatory hot topics emerge is increasing over time,” Hildyard said. Additionally, trade surveillers’ “goal should be to ‘create’ the next big scandal and make sure it doesn’t happen on their watch, in their bank. That requires that they understand behaviors they weren’t previously monitoring for.”


Transforming Text and Data Into a True Knowledge Base
One of the steps in text mining is “relationship identification.” Once entities are identified and enriched, they are connected to other entities; for example, “Foggy Bottom is in Washington, DC”, “Foggy Bottom is near The White House” and “Foggy Bottom is east of Georgetown.” What just happened? We used Open Linked Data (LOD) to verify Foggy Bottom as a neighborhood that exists in Washington DC while also connecting it to other entities. LOD knows that DC is a “District” (not a state) and that it is within the United States. Preexisting facts were combined with results from text analysis to expand the knowledge base.


APIs with Swagger : An Interview with Reverb’s Tony Tam
First, we don’t want to try to stuff every possible feature inside the specification itself. Early on, someone brought up embedding rate-limiting information into the spec. But it would be very difficult to generalize, and would pollute the spec over a feature that possibly many people wouldn’t care about. Next, one thing we learned through the initial versions of Swagger is that it’s easy to write invalid specifications without a simple and robust validator. We chose to use JSON Schema validations, and even built it directly into Swagger-UI. It is an important part of the tooling to help developers write valid Swagger definitions. Removing structural constraints from the spec AND having a robust validation tool would be very difficult.


Case study: What the enterprise can learn from Etsy's DevOps strategy
“You have to be able to demonstrate to the larger business why it’s not just a buzzword and can add value to the business, and the only way to do that is to give them a concrete project and show them how it has positively affected the business,” he says. “The people who make the decisions at the top of the pile may be more business-minded than technically so, and you need to speak their language and demonstrate the impact it has had on key performance indicators or revenue that quarter. “You need to sell the idea to them in business terms because IT and development are service organisations that exist to fulfil the priorities of the business,” Cowie adds.


A Brief History of Big Data Everyone Should Read
Long before computers (as we know them today) were commonplace, the idea that we were creating an ever-expanding body of knowledge ripe for analysis was popular in academia. Although it might be easy to forget, our increasing ability to store and analyze information has been a gradual evolution – although things certainly sped up at the end of the last century, with the invention of digital storage and the internet. With Big Data poised to go mainstream this year, here’s a brief(ish) look at the long history of thought and innovation which have led us to the dawn of the data age.



Quote for the day:

"Every leader needs to look back once in awhile to make sure he has followers." -- Kouzes and Posner