Championing privacy-first security: Harmonizing privacy and security compliance
When security solutions are crafted with privacy as a central consideration,
organisations can deploy robust security measures while safeguarding the
personal data of their customers and employees. A comprehensive cost-benefit
analysis reveals significant advantages in adopting a privacy-first approach to
security. For instance, proactively blocking malware before it infiltrates an
organisation’s systems can avert a potential data breach. Given the average cost
of US$4.45 million in 2023, coupled with the consequential impact on brand
reputation and legal ramifications, preventing even a single data breach becomes
paramount for any company. Hence, the importance of industry-leading security
measures is indisputable. Any reputable security company should provide
solutions that limit its access to sensitive data and ensure the protection of
the personal data entrusted to its care. ... A privacy-first security program
assesses the risks associated with both implementing and not implementing
security measures. If the advantages of deploying a security solution, such as
email scanning, outweigh the drawbacks – which is highly probable – the
organisation should proceed with the careful implementation of this capability.
Far Memory Unleashed: What Is Far Memory?
Far memory is a memory tier between DRAM and Flash that has a lower cost per
GB than DRAM and a higher performance than Flash. Far memory works by
disaggregating memory and allowing nodes or machines to access the memory of a
remote node/machine via compute express link. Memory is the most contested and
least elastic resource in a data center. Currently, servers can only use local
memory, which may be scarce on the local system but abundant on other
underutilized servers. With far memory, local machines can use remote
machine’s memory. By introducing far memory into the memory tier and moving
less frequently accessed data to far memory, the system can perform
efficiently with low DRAM and reduce the total cost of ownership. Far memory
uses a remote machine’s memory as a swap device, either by using idle machines
or by building memory appliances that only serve to provide a pool of memory
shared by many servers. This approach optimizes memory usage and reduces
over-provisioning. However, far memory also has its own challenges. Swapping
out memory pages to remote machines increases the failure domain of each
machine, which can lead to a catastrophic failure of the entire cluster.
Four reasons your agency's security infrastructure isn't agile enough
There are four key considerations for integrating security architecture
effectively in an agile environment - Cross-Functional Collaboration: Security
experts must actively engage with developers, testers, and product owners.
Collaborating with experts helps create a shared understanding of security
requirements and facilitates quick resolution of security-related issues.
Embedding security professionals within Agile teams can enhance real-time
collaboration and ensure consistent security controls. Security Training
and Awareness: Given the rapid pace of an Agile sprint, all team members
should be equipped with the knowledge to write secure code. ... Foster a
Security Culture: Foster a culture where security is seen as everyone's
responsibility, not just the security team's. Adapt the organizational mindset
to value security equally with other business objectives. ... Security
Champions within Agile Teams: Identify and nurture 'Security Champions' within
each Agile team. These individuals with a keen interest in security act as a
bridge between the security team and their respective agile teams. They help
promote security best practices, ensuring security is not overlooked amidst
other technical considerations.
AI Legislation: Enterprise Architecture Guide to Compliance
Artificial intelligence (AI) tools are so easy to leverage that they can be
used by anyone within your organization without technical support. This means
that you need to keep a careful eye on, not just the authorized applications
you leverage, but what AI tools your colleagues could be using without
authorization. In leveraging AI tools to generate content for your
organization, your employees could unwittingly input private data into the
public instance of ChatGPT. Not only does this share that data with ChatGPT's
vendor, OpenAI, but it actually trains ChatGPT on that content, meaning the AI
tool could potentially output that information to another user outside of your
organization. Alternatively, overuse of generative AI tools without proper
supervision could lead to factual or textual errors being published to your
customers. Gen AI tools need careful supervision to ensure they don't
"hallucinate" or produce mistakes, as they are unable to self-edit. It's
equally important to be able to report back to legislators on what AI is being
used across your company, so they can see you're compliant. This will likely
become a regulatory requirement in the near future.
Choosing a disaster recovery site
The first option is to set up your own secondary DR data center in a different
location from your primary site. Many large enterprises go this route; they
build out DR infrastructure that mirrors what they have in production so that,
at least in theory, it can take over instantly. The appeal here lies in
control. Since you own and operate the hardware, you dictate compatibility,
capacity, security controls and every other aspect. You’re not relying on any
third party. The downside of course, lies in cost. All of that redundant
infrastructure sitting idle doesn’t come cheap. ... The second approach is to
engage an external DR service provider to furnish and manage a recovery site
on your behalf. Companies like SunGard built their business around this model.
The appeal lies in offloading responsibility. Rather than build out your own
infrastructure, you essentially reserve DR data center capacity with the
provider. ... The third option for housing your DR infrastructure is
leveraging the public cloud. Market leaders like AWS and Azure offer seemingly
limitless capacity that can scale to meet even huge demands when disaster
strikes.
How CISOs navigate policies and access across enterprises
Simply speaking, if existing network controls are now being moved to the
cloud, the scope of technical controls does not drastically differ from legacy
approaches. The technology, however, has massively evolved towards
platform-centric controls, and that for a good reason. Isolated controls cause
complexity, and if you are moving your perimeter to a hyperscaler, both your
users and their devices will no longer be managed by the corporate on-prem
security controls either. A good CASB to broker between user and data is key,
as is identity and access management. What’s now new is workload protection
requirements รก la CSAP technology. In addition to increasing sophistication
and the number of security threats and successful breaches, most enterprises
further increase risk by “rouge IT” teams leveraging cloud environments
without the awareness and management by security teams. Cloud deployments are
typically deployed faster and with less planning and oversight than data
center or on-site environment deployments. Cloud security tools should be
an extension of your other premise-based tools for ease of management,
consistency of policy enforcement and cost savings due to additional purchase
commitments, training, and certification non-duplicity.
What to Know About Machine Customers
In the realm of customer service and support, machine customers are like
virtual assistants or smart devices (think of Siri or Alexa) that carry out
customer service tasks on behalf of actual human customers. Alok Kulkarni, CEO
and Co-founder of Cyara, says the emergence of machine customers introduces a
new dynamic, requiring organizations to adapt their existing support
strategies. “This might include developing specific interfaces and
communication channels tailored for interactions with machine customers,” he
explains in an email interview. Organizations must create additional
self-service options specifically designed for machine customers. “Unlike
traditional customer support approaches, catering to machine customers
requires a nuanced understanding of their specific needs and operational
dynamics,” Kulkarni explains. This means designing self-service interfaces
that are not only user-friendly for machines but also align with the
intricacies of autonomous negotiation and purchasing processes. These
interfaces should empower machine customers to navigate through various stages
of transactions autonomously, from product selection to payment processing,
ensuring a streamlined and frictionless experience.
Google: Govs Drive Sharp Growth of Commercial Spyware Cos
Much of the concern has to do with the explosion in the availability of tools
and services that allow governments and law enforcement to break into target
devices with impunity, harvest information from them, and spy unchecked on
victims. The vendors selling these tools — most of which are designed for
mobile devices — have often openly pitched their wares as legitimate tools
that aid in law enforcement and counter-terrorism efforts. But the reality is
that repressive governments have routinely used spyware tools against
journalists, activists, dissidents, and opposition party politicians, said
Google. The company's report cites three instances of such misuse: one that
targeted a human rights defender working with a Mexico-based rights
organization; another against an exiled Russian journalist; and the third
against the co-founder and director of a Salvadorian investigative news
outlet. The researcher attributes much of the recent growth in the CSV market
to strong demand from governments around the world to outsource their need for
spyware tools rather than have an advanced persistent threat build them
in-house.
How To Build Autonomous Agents – The End-Goal for Generative AI
From a technology perspective, there are five elements that go into autonomous
agent designs: the agent itself, for processing; tools, for interaction;
prompt recipes, for prompting and planning; memory and context, for training
and storing data; and APIs / user interfaces, for interaction. The agent at
the center of this infrastructure leverages one or more LLMs and the
integrations with other services. You can build this integration framework
yourself, or you can bring in one of the existing orchestration frameworks
that have been created, such as LangChain or LlamaIndex. The framework should
provide the low-level foundational model APIs that your service will support.
It connects your agent to the resources that you will use as part of your
overall agent, including everything from existing databases and external APIs,
to other elements over time. It also has to take into account what use cases
you intend to deliver with your agent, from chatbots to more complex
autonomous tasks. Existing orchestration frameworks can take care of a lot of
the heavy lifting involved in managing LLMs, which makes it much easier and
faster to build applications or services that use GenAI.
How Platform and Site Reliability Engineering Are Evolving DevOps
Actually, failure should not just be OK but welcome. Most organizations are
averse to failure, but it’s only through our failures in these spaces that we
can learn and grow and figure out how to best position, leverage, and continue
to imagine the roles of DevOps, platform engineers, and SRE. I’ve seen this
play out in large companies that went all in on DevOps and then realized that
they needed a team focused on breaking down any barriers that presented
themselves to developers. At scale, DevOps - even with the tools provided by
the internally focused platform engineering team - didn’t really cut it. These
companies then integrated the SRE function, which filled DevOps’ reliability
and scalability gaps. That worked until these companies realized that they
were reinventing the wheel - dozens of times. Different engineering teams
within the organization were doing things just differently enough - different
setups, different processes, different expectations - that they needed
separate setups to put out a service. The SREs were seeing all of this after
the fact, which led them to circle back to the realization that different
teams needed to be using the same development building blocks. Frustrating?
Yes. The cost of increasing efficiency in the future? Absolutely.
Quote for the day:
“It’s better to look ahead and
prepare, than to look back and regret.” -- Jackie Joyner-Kersee
No comments:
Post a Comment