Beyond the breach: How cloud ransomware is redefining cyber threats in 2024
Unlike conventional ransomware that targets individual computers or on-premises
servers, attackers are now setting their sights on cloud infrastructures that
host vast amounts of data and critical services. This evolution represents a new
frontier in cyber threats, requiring Indian cybersecurity practitioners to
rethink and relearn defence strategies. Traditional security measures and last
year’s playbooks are no longer sufficient. Attackers are exploiting
misconfigured or poorly secured cloud storage platforms such as Amazon Web
Services (AWS) Simple Storage Service (S3) and Microsoft Azure Blob Storage. By
identifying cloud storage buckets with overly permissive access controls,
cybercriminals gain unauthorised entry, copy data to their own servers, encrypt
or delete the original files, and then demand a ransom for their
return. ... Collaboration and adaptability are essential. By understanding
the unique challenges posed by cloud security, Indian organisations can
implement comprehensive strategies that not only protect against current threats
but also anticipate future ones. Proactive measures—such as strengthening access
controls, adopting advanced threat detection technologies, training employees,
and staying informed—are crucial steps in defending against these evolving
attacks.
Harnessing AI’s Potential to Transform Payment Processing
There are many use cases that show how AI increases the speed and convenience
of payment processing. For instance, Apple Pay now offers biometric
authentication, which uses AI facial recognition and fingerprint scanning to
authenticate users. This enables mobile payment customers to use quick and
secure authentication without remembering passwords or PINs. Similarly, Apple
Pay’s competitor, PayPal, uses AI for real-time fraud detection, employing ML
algorithms to monitor transactions for signs of fraud and ensure that
customers’ financial information remains secure. ... One issue is AI systems
rely on massive amounts of data, including sensitive data, which can lead to
data breaches, identity theft, and compliance issues. In addition, AI
algorithms trained on biased data can perpetuate those biases. Making matters
worse, many AI systems lack transparency, so the bias may grow and lead to
unequal access to financial services. Another issue is the potential
dependence on outside vendors, which is common with many AI technologies. ...
To reduce the current risks associated with AI and safely unleash its full
potential to improve payment processing, it is imperative for organizations to
take a multi-layered approach that includes technical safeguards,
organizational policies, and regulatory compliance.
Do you need an AI ethicist?
The goal of advising on ethics is not to create a service desk model, where
colleagues or clients always have to come back to the ethicist for additional
guidance. Ethicists generally aim for their stakeholders to achieve some level
of independence. “We really want to make our partners self-sufficient. We want
to teach them to do this work on their own,” Sample said. Ethicists can
promote ethics as a core company value, no different from teamwork, agility,
or innovation. Key to this transformation is an understanding of the
organization’s goal in implementing AI. “If we believe that artificial
intelligence is going to transform business models…then it becomes incumbent
on an organization to make sure that the senior executives and the board never
become disconnected from what AI is doing for or to their organization,
workforce, or customers,” Menachemson said. This alignment may be especially
necessary in an environment where companies are diving head-first into AI
without any clear strategic direction, simply because the technology is in
vogue. A dedicated ethicist or team could address one of the most foundational
issues surrounding AI, notes Gartner’s Willemsen. One of the most frequently
asked questions at a board level, regardless of the project at hand, is
whether the company can use AI for it, he said.
Why We Need Inclusive Data Governance in the Age of AI
Inclusive data governance processes involve multiple stakeholders, giving
equal space in this decision making to diverse groups from civil society, as
well as space for direct representation of affected communities as active
stakeholders. This links to, but is an idea broader than, the concept of
multi-stakeholder governance for technology, which first came to prominence at
the international level, in institutions such as the Internet Corporation for
Assigned Names and Numbers and the Internet Governance Forum. ... Involving
the public and civil society in decisions about data is not cost-free. Taking
the steps that are needed to surmount the practical challenges, and skepticism
about the utility of public involvement in a technical and technocratic field,
frequently requires arguments that go beyond it being the right thing to do.
... The risks for people, communities and society, but also for organizations
operating within the data and AI marketplace and supply chain, can be reduced
through greater inclusion earlier in the design process. But organizational
self-interest will not motivate the scope or depth that is required. Reducing
the reality and perception of “participation-washing” means requirements for
consultation in the design of data and AI systems need to be robust and
enforceable.
Strategies to navigate the pitfalls of cloud costs
If cloud customers spend too much money, it’s usually because they created
cost-ineffective deployments. It’s common knowledge that many enterprises
“lifted and shifted” their way to the clouds with little thought about how
inefficient those systems would be in the new infrastructure. ... Purposely or
not, public cloud providers created intricate pricing structures that are
nearly incomprehensible to anyone who does not spend each day creating cloud
pricing structures to cover every possible use. As a result, enterprises often
face unexpected expenses. Many of my clients frequently complain that they
have no idea how to manage their cloud bills because they don’t know what
they’re paying for. ... Cloud providers often encourage enterprises to
overprovision resources “just in case.” Enterprises still pay for that unused
capacity, so the misalignment dramatically elevates costs without adding
business value. When I ask my clients why they provision so much more storage
or computing resources beyond what their workload requires, the most common
answer is, “My cloud provider told me to.” ... One of the best features of
public cloud computing is autoscaling so you’ll never run out of resources or
suffer from bad performance due to insufficient resource provisioning.
However, autoscaling often leads to colossal cloud bills because it often is
triggered without good governance or purpose.
Your IT Team Isn't Ready For Change Management If They Can't Answer These 3 Questions
Testing software before you encounter failure rates is key, but never should
you be exposed to failure rates with this level of real world impact. Whether
it’s due to third party systems or the companies themselves, their brand will
be the one in tatters due to the end customer experience. Enter Change
Management and the possibility for, if done right, the prevention of these
kinds of enormous IT failures. ... The ever-evolving nature of technology,
including cloud scaling, infrastructure as code, and frequent updates such as
‘Patch Tuesday’ means that organisations must constantly adapt to change.
However, this constant change introduces challenges such as “drift”—a term
that refers to the unplanned deviations from standard configurations or
expected states within an IT environment. Think of it like a pesky monkey in
the machine. Drift can occur subtly and often goes unnoticed until it causes
significant disruptions. It also increases uncertainty and doubt in the
organisation making Change Management and Release Management harder, creating
difficulties to plan and execute changes safely. ... To be effective, Change
Management needs to be able to detect and understand drift in the environment
to have a full understanding of Current State, Risk Assessment and Expected
Outcomes.
RIP Open Core — Long Live Open Source
Open-core was originally popular because it allowed companies to build a
community around a free product version while charging for a more full,
enterprise-grade version. This setup thrived in the 2010s, helping companies
like MongoDB and Redis gain traction. But times have changed, and today,
instead of enhancing a company’s standing, open-core models often create more
problems than they solve. ... While open-core and source-available models had
their moment, companies are beginning to realize the importance of true open
source values and are finding their way back. This return to open source is a
sign of growth, with businesses realigning with the collaborative spirit at
the core (wink) of the OSS community. More companies are adopting models that
genuinely prioritize community engagement and transparency rather than using
them as marketing or growth tactics. ... As the open-core model fades, we’re
seeing a more sustainable approach take shape: the Open-Foundation model. This
model allows the open-source offering to be the backbone of a commercial
offering without compromising the integrity of the OSS project. Rather, it
reinforces it as a valuable, standalone product that supports the commercial
offering instead of competing against it.
Why SaaS Backup Matters: Protecting Data Beyond Vendor Guarantees
Most IT departments have long recognized the importance of backup and recovery
for applications and data that they host themselves. When no one else is
managing your workloads and backing them up, having a recovery plan to restore
them if necessary is essential for minimizing the risk that a failure could
disrupt business operations. But when it comes to SaaS, IT operations teams
sometimes think in different terms. That's because SaaS applications are
hosted and managed by external vendors, not the IT departments of the
businesses that use SaaS apps. In many cases, SaaS vendors provide uptime or
availability guarantees. They don't typically offer details about exactly how
they back up applications and data or how they'll recover data in the event of
a failure, but a backup guarantee is typically implicit in SaaS products. ...
Historically, SaaS apps haven't featured prominently, if at all, in backup and
recovery strategies. But the growing reliance on SaaS apps — combined with the
many risks that can befall SaaS application data even if the SaaS vendor
provides its own backup or availability guarantees — makes it critical to
integrate SaaS apps into backup and recovery plans. The risks of not backing
up SaaS have simply become too great.
Biometrics in the Cyber World
Biometrics is known to strengthen security in many ways. Some ways this could
include stronger authentication, user convenience, and reduction of risk of
identity theft. With the uniqueness of the user, this adds a layer of security
to authentication. Traditional authentication, such as passwords, contains
weak combinations that can easily be breached, so using biometrics can prevent
that. People constantly forget their passwords and always end up resetting
them. Since biometrics is entirely connected with one’s identity, it will no
longer be an inconvenience to forget a password. Since it is very usual for
hackers to attempt to get into an account by guessing the password, biometrics
does not accommodate this since a hacker cannot guess a password when the
uniqueness of one’s identity is what is sought for authentication. This is one
of the pros of biometric systems, which should reduce the risk of identity
theft. Some challenges that biometrics could have include privacy concerns,
false positives and negatives, and bias. Since it is personal information,
biometric data can lead to privacy violations. Storage of personal data, which
is sensitive, needs to be following the regulations on privacy such as GDPR
and CCPA.
Data Architectures in the AI Era: Key Strategies and Insights
Accelerating results in data architecture initiatives can be achieved in a
much quicker fashion if you start with the minimum needed and build from there
for your data storage. Begin by considering all use cases and finding the one
component needed to develop so a data product can be delivered. Expansion can
happen over time with use and feedback, which will actually create a more
tailored and desirable product. ... Educating your key personnel on the
importance of being able and ready to make the shift from previously familiar
legacy data systems to modern architectures like data lakehouses or hybrid
cloud platforms. Migration to a unified, hybrid, or cloud-based data
management system may seem challenging initially, but it is essential for
enabling comprehensive data lifecycle management and AI-readiness. By
investing in continuous education and training, organizations can enhance data
literacy, simplify processes, and improve long-term data governance,
positioning themselves for scalable and secure analytics practices. ... By
being prepared for the typical challenges of AI, problems can be predicted and
anticipated which can help to reduce downtime and frustration in the
modernization of data architecture.
Quote for the day:
“The final test of a leader is that he
leaves behind him in other men the conviction and the will to carry on.” –
Walter Lippmann
No comments:
Post a Comment