Daily Tech Digest - October 06, 2023

Cloud infrastructure spending is growing

Although I love to be right about the strong cloud spending, that does not mean it’s suitable for all enterprises. Indeed, the trend will be to overspend, even after net-new finops deployments that closely monitor where the dollars are spent. We must focus on accountability, automation, and discipline around allocating and paying for cloud resources. I suspect many cloud deployments are hugely underoptimized and need a tune-up. Even though some of this shared infrastructure spending is unavoidable, CIOs need to review how the spending occurs and look for opportunities to save dollars without reducing the value generated by these systems. I suggest companies consider all other options, such as bringing some processing into enterprise data centers. Those prices have been falling while they have been stable or rising on the public cloud side. Also, many systems function in isolation and don’t benefit much from existing within a public cloud. Simple storage is one example, and many enterprises are putting those systems on-premises these days.


BAs are responsible for creating new models that support business decisions by working closely with finance and IT teams to establish initiatives and strategies aimed at improving revenue and/or optimizing costs. Business analysts need a “strong understanding of regulatory and reporting requirements as well as plenty of experience in forecasting, budgeting, and financial analysis combined with understanding of key performance indicators,” according to Robert Half Technology. ... Business analysts need to know how to pull, analyze and report data trends, share that information with others, and apply it to business goals and needs. Not all business analysts need a background in IT if they have a general understanding of how systems, products, and tools work. Alternatively, some business analysts have a strong IT background and less experience in business but are interested in shifting away from IT into this hybrid role. The role often acts as a communicator between the business and IT sides of the organization, so having extensive experience in either area can be beneficial for business analysts.


AI Needs Data More Than Data Needs AI

While data plays a foundational role in AI, the reverse is not true. Data doesn't inherently need AI to exist or be valuable. Data, in various forms, has been collected and analyzed for centuries without the need for sophisticated AI algorithms. Data on its own can provide valuable insights and inform decision-making processes. Therefore, organizations should not blindly chase the AI hype at the cost of ignoring the importance of data management and data quality. The role of AI is to take the computation and insights of good quality data to the next level and not necessarily attempt to fix the decades-old data management processes. ... While AI relies heavily on data for its operation and evolution, data can benefit from AI in several ways. Data Management: AI can help automate data management tasks, making it easier to process, clean and organize large datasets. Predictive Insights: AI can uncover patterns and insights in data that may not be immediately apparent to humans, enhancing the value of the data.


Enterprises see AI as a worthwhile investment

Despite prior industry research indicating that 90% of AI initiatives fail to produce substantial ROI and roughly half never leave the prototype stage, the overwhelming majority of respondents to this survey (92%) find business value from their models in production and 66% feel their models have delivered results that are outstanding or exceed expectations. Common use cases for AI among these leading-edge organizations include personalizing the customer experience, fraud detection, optimizing sales and marketing and improving real-time decision making. Their success of this group offers a basic roadmap that other organizations should consider when developing their own best practices, including: Approach: A majority of responding organizations have a robust, defined approach and a dedicated team for monitoring ML models in production. In fact among larger enterprises, 71% have at least 100 people working in ML while over half have more than 250. 


5 Strategies for Cloud Security in Health Care

Adopting data security in the cloud doesn’t mean merely uploading patient data to S3 and enabling encryption. There are many security controls that need to be in place before a single patient record is migrated. For instance, there is particular concern about data security on medical devices and wireless body area networks (devices that are embedded in a patient’s body). Obviously, it’s vital to secure such devices from exploits. When running services on the cloud, you should review all relevant data privacy considerations and encryption controls, including data encryption, public-key encryption, identity-based encryption, identity-based broadcast encryption and attribute-based encryption. Then adopt a framework for achieving secure and controlled identity access using federation (like OpenID Connect, which is not the same as OpenID, or SAML). Finally, you should ensure that monitoring and audit controls are in place to maintain confidentiality. You should also have an incident response plan in place to handle crisis scenarios in the event of an incident. 


Financial Institutions Turn to AI and Cloud to Solve Data Challenges

In data management, the potential uses of GenAI, powered by large language models, has been recognised by many financial institutions, including State Street. For instance, it can help in the cross-mapping of datasets, the classifying of data and more generalist applications such as summarising reports and responding to plain English inquiries. ... The Alpha platform uses GenAI with Snowflake as a strategic partner providing the data foundation of the platform. Snowflake’s cloud-native architecture streamlines data sharing and governance, enables faster time to market for data-centric applications, and offers a rich environment of AI and machine learning-based capabilities for data scientists, quants and engineers. “Every few years, the technology landscape re-sets, creating a small window of opportunity that in turn enables a giant leap in innovation; GenAI is the opportunity that will define the new set of industry leaders over the next decade,” State Street Executive Vice President and Chief Architect Aman Thind tells A-Team Group.


Building data center networks for GenAI fabric enablement

Building GenAI data centers from a network perspective differs greatly from traditional data center buildouts -- or even those that were designed to support high-performance computing (HPC). ... After all, the pace of a GenAI application is only as fast as its slowest component. If properly built, the network can be eliminated as a potential performance bottleneck. Building a highly scalable network is also key to GenAI data centers as it enables future growth capacity. Network switch fabrics must include hardware that can expand horizontally and vertically, as well as use network OSes on switching hardware that include advanced features, such as packet spraying, load awareness and intelligent traffic redirection. These features provide automated rerouting of traffic within the network and between GPU processing units that may become overloaded. ... Early GenAI adopters have concluded that the use of multisite or micro data centers is the best option to accommodate this level of density. And, yet again, this puts pressure on the network interconnecting these sites to be as high-performing and resilient as possible.


Breach Roundup: Still Too Much ICS Exposed on the Internet

Apple responded to an actively exploited zero-day flaw in iOS and iPadOS on Wednesday with the release of security patches. The identified vulnerability, tracked as CVE-2023-42824, exists in the kernel and may allow an attacker to elevate privileges. "Apple is aware of a report that this issue may have been actively exploited against versions of iOS before iOS 16.6," the company said. The update also addresses CVE-2023-5217, a WebRTC component issue. WebRTC is an open-source project that supports real-time computing between browsers and mobile applications, powering uses such as video and voice calling. ... Sony Interactive Entertainment alerted around 6,800 individuals about a cybersecurity breach. The intrusion resulted from an unauthorized party exploiting a zero-day vulnerability, tracked as CVE-2023-34362, in the MOVEit file transfer platform. This critical-severity SQL injection flaw, leading to remote code execution, was used by the Clop ransomware gang in widespread attacks in late May. 


8 Ways to Combat Ageism in Your Job Search

Workplace experts say candidates can combat this by showing what efforts they've made to quickly pick up new skills and show enthusiasm for future learning. That might mean enrolling in extra training courses, getting new certifications and highlighting them in your résumé or interview, North said. Younger workers may need to show that they have taken proactive measures to learn new job skills they may lack. Older workers may want to show that they can keep up with fast-paced environments and various tech tools. ... "If you don't have to input this information, don't volunteer it," he said, adding that phrases like 40-plus years of experience also may not be best. Instead, stick to your skills and experiences. If you lack experience in one area, show how your skills are transferrable for this specific job. You can also be clear about any kind of transition, like a career change, or gap in employment by placing it in an executive summary section at the top of your résumé, Freeman said. Quantify your previous work's impact with numbers or qualify it by explaining how it affected the results.


Ransomware Crisis, Recession Fears Leave CISOs in Tough Spot

With a new ransomware target being attacked every 14 seconds, organizations must prioritize ransomware prevention. With its developing sophistication, mitigating ransomware is increasingly more challenging. There's no silver bullet to eradicate attacks, and having to operate in a tight market adds a layer of complexity. CISOs and security leaders must focus on the best return on investment while building out a multilayered approach for improving their overall IT security. One strategy to accomplish this is managing attack vectors using encrypted channels with preventive technologies that can stop adversaries before they have a chance to compromise networks or while they are executing their multistep campaigns. ... Ransomware gangs also take advantage of legitimate websites encrypted with SSL/TLS to look secure, but have been infected with drive-by downloads. And cybercriminals leech onto browser vulnerabilities that can lead to infection when the entry point is encrypted, allowing encrypted threats embedded with malicious payloads to go unnoticed.



Quote for the day:

“People are not lazy. They simply have important goals – that is, goals that do not inspire them.” -- Tony Robbins

Daily Tech Digest - October 05, 2023

AI and Overcoming User Resistance

If users are concerned, and even worried about AI, it could lead to user resistance, which is a dynamic that IT pros are familiar with from their history of implementing new systems that alter business processes, require employee retraining, and may even change employee jobs. So, are process change and user resistance any different when you introduce AI? I would argue yes. You’re not just retraining an employee on a new set of steps for processing an invoice or taking an order. You’re actually introducing an automated thinking process into what an employee has been doing. Now, technology is going to make or recommend decisions that the employee used to make. This can lead to employees experiencing a loss of empowerment and control. ... This is exactly the “sweet spot” that companies (and IT) should aim for with AI projects: an environment where everyone sees beneficial value from AI, and where no one feels disenfranchised. This is an achievable environment if users are engaged early in business process redefinition and in how AI will work. 


Eyes everywhere: How to safely navigate the IoT video revolution

Users are rightfully wary of bringing even more cameras into their homes and offices. The good news is that they, too, can protect their camera-enabled devices with some simple steps. First, customize. This includes changing default usernames and passwords, updating the device’s firmware and software, and staying informed about the latest security threats. This is a simple yet effective way to create a barrier between yourself and would-be hackers. Next, take it to the edge. Processing and storing data at the edge instead of the cloud is another surefire way to protect your endpoints. After all, by storing the information under your own lock and key, you can be sure about who can access it and how. Users also benefit from reduced latency by storing the information closer to home, which is particularly important with heavy video feeds. Finally, buy trusted brands. Attack surfaces are only as strong as their weakest link. So, chose companies that have a proven track record when it comes to privacy and security. 


Why HTTP Caching Matters for APIs

In some caching strategies, especially for dynamic resources, the cache can store not only the complete response but also the individual elements or changes that make up the response. This approach is known as “delta caching” or “incremental caching.” Instead of sending the complete response, delta caching sends only the changes or updates made to the cached version of the resource. ... Delta caching is particularly useful for scenarios where resources change frequently, but the changes are relatively small compared to the complete resource. For example, in a collaborative document editing application, delta caching can be employed to send only the changes made by a user to a shared document, instead of sending the entire document every time it is updated. ... Caching enhances application resilience by reducing the risk of service disruptions during periods of high demand. By serving cached responses, even if the backend servers experience temporary performance issues, the application can continue to respond to a significant portion of requests from the cache. The caching layer acts as a buffer between the backend servers and the clients.


Author Talks: How to speak confidently when you’re put on the spot

People become nervous for many reasons. More than 75 percent of people report being nervous in high-stakes communication, be it planned or spontaneous. Past experience could be a factor, as well as high stakes and the importance of the goals you’re trying to achieve. Those of us who study this at an academic level believe that the nervousness is wired into being human. We see this across all cultures. We see it develop typically in the early teen years and progress from there. There’s an evolutionary component to it. One of the most helpful tips is normalizing the anxiety that you feel. You’re not alone. ... My anxiety management plan has three steps. The first thing I do is hold something cold in the palms of my hand before I speak. That cools me down. Secondly, I say tongue twisters to warm up my voice and also to get myself in the moment. Third, I remind myself, “I am in service of my audience. I am here to help them.” That really gets me other-focused rather than self-focused. That’s my anxiety management plan. I encourage everybody to find a plan that works for them.


Dell customizes GenAI and focuses on data lakehouse

Being able to fine tune as well as train generative AI is a process that relies on data, lots and lots of data. For enterprise use cases, that data isn’t just generic data taken from a public source, but rather is data that an organization already has in its data centers or cloud deployments and is likely also spread across multiple locations. To help enable enterprises to fully benefit from data for generative AI, Dell is building out an open data lakehouse platform. The data lakehouse concept is one that was originally pioneered by Databricks, as a way of enabling organizations to more easily query data stored in cloud object storage based data lakes. The Dell approach is a bit more nuanced in that it is taking a hybrid approach to data, with a goal of being able to query data across on-premises as well as mutli-cloud deployments. Greg Findlen, senior VP data management at Dell explained during the press briefing that the open data lakehouse will be able to use Dell storage and compute capabilities as well as multi-cloud storage. 


Don’t try running with data before you can walk

In South Africa, data governance tends to be a grudge investment based on regulatory issues. However, organisations that don’t do the basics well, and don’t have mature data governance and established frameworks in place, may well find they are spending on analytics technologies that don’t live up to expectations. What stands in the way of getting governance right? Firstly, it’s not easy. It involves all stakeholders across all domains. It may require a mindset change, and users may need to learn to use new technology. Secondly, it can be expensive, and it may take time before the organisation sees the value of it. One of the biggest problems is that the value of data governance investments is difficult to quantify in monetary terms. ... Data products should be supported by the entire CDO capability – including the CDO, data owners and data stewards – as well as IT, to ensure the data products will add the required business value. Owners and stewards need to identify and curate the required data for the products, while also ensuring good quality data and metadata management to make it more usable for broader business.


Yes, Software Development is an Assembly Line, but not Like That

Manufacturing engineers produce assembly lines and manufacturing processes that can produce those units of value. Software engineers are largely the same, also producing systems and processes that deliver units of value. The manufactured widget of software is actually the discrete user interactions with those features and pieces of software, not the features themselves. The assembly line in software engineering isn’t, as many think, the engineers producing features. ... Systems like Total Quality Management, which are focused on driving a cultural mindset of continuous improvement and an entire company focused on providing very low defect rates, easily translate to customer satisfaction in software organizations. Just to pick on TQM a bit, if we were to adapt it to software, we would focus on the number of times users are impacted by a defect more than the number of open bugs. Instead of tracking the number of defects and searching for more, we would be tracking the number of users who either failed to receive the promised value from the product or had severely diminished value.


Cloud Services Without Servers: What’s Behind It

“The basic idea of serverless computing has been around since the beginning of cloud computing. However, it has not become widely accepted,” explains Samuel Kounev, who heads the JMU Chair of Computer Science II (Software Engineering). But a shift can currently be observed in the industry and in science, the focus is increasingly moving towards serverless computing. A recent article in the Communications of the ACM magazine of the Association for Computing Machinery (ACM) deals with the history, status and potential of serverless computing. Among the authors are Samuel Kounev and Dr. Nikolas Herbst, who heads the JMU research group “Data Analytics Clouds”. ... “NoOps” is the first, which stands for “no operations”. This means, as described above, that the technical server management, including the hardware and software layers, is completely in the responsibility of the cloud provider. The second principle is “utilisation-based billing”, which means that only the time during which the customer actively uses the allocated computing resources is billed. 


7 sins of software development

Some software development issues can be fixed later. Building an application that scales efficiently to handle millions or billions of events isn’t one of them. Creating effective code with no bottlenecks that surprise everyone when the app finally runs at full scale requires plenty of forethought and high-level leadership. It’s not something that can be fixed later with a bit of targeted coding and virtual duct tape. The algorithms and data structures need to be planned from the beginning. That means the architects and the management layer need to think carefully about the data that will be stored and processed for each user. When a million or a billion users show up, which layer does the flood of information overwhelm? How can we plan ahead for those moments? Sometimes this architectural forethought means killing some great ideas. Sometimes the management layer needs to weigh the benefits with the costs of delivering a feature at scale. Some data analysis just doesn’t work well at large scale. Some formulas grow exponentially with more users. 


Organizations grapple with detection and response despite rising security budgets

For better understanding and evaluation, the study was able to categorize the responding organizations into "secure creators" and "prone enterprises." The grouping was done on the basis of the number of solutions used, the adoption of emerging technologies, and the use of technologies to simplify their automation environments. The study found that secure creators are more satisfied with their approach to cybersecurity, experience fewer cybersecurity incidents, and can detect and respond to incidents quicker. About 70% of them are early adopters of emerging technologies. The secure creators are also more focused on extracting the most value from specific advanced solutions, with 62% already using or in the late stages of implementing AI/ML solutions, as compared to only 45% of the prone enterprises. "When it comes to technology, the more clutter an organization has in its armory, the harder it is to pick up signals and get on top of issues quickly," Watson said.



Quote for the day:

"You’ll never achieve real success unless you like what you’re doing." -- Dale Carnegie

Daily Tech Digest - October 04, 2023

The Big Threat to AI: Looming Disruptions

As if semiconductor supply chain issues weren’t enough of a problem for AI production, other supply chains are piling on the challenges. "AI is software and open-source code makes up 90% of most codebases, which means the open source software supply chain has just as much, if not more, impact on AI production than regulated hardware components,” says Feross Aboukhadijeh, founder and CEO of Socket. The impact is potentially widespread given there are many open source AI models and tools on the market today and more are coming. ... There are numerous efforts afoot to relieve these concerns and secure a prime slice of the AI market pie. For what corporation does not envy Nvidia right now? “Many countries are trying to increase their piece of the global supply chain capacity and/or to onshore as much as possible through subsidies and other incentives. This has spurred significant investment and activity, but it remains to be seen whether these investments will address the supply chain problems in a timely or appropriate manner,” says Almassy.


When to Scale and When Not to Scale

Scaling is a nuanced decision in the agile journey, bridging the demands of complexity and rapid market needs. While the lure of scaling promises greater coordination, efficient handling of product intricacies, and swifter market responses, it's pivotal to approach it judiciously. It's not just about expanding teams or implementing frameworks; it's about recognizing when the product's complexity or market dynamics truly warrant a scaled approach. On the flip side, scaling without a clear strategy can introduce unforeseen challenges. From the inadvertent hiring of too many junior roles to the formation of functional silos, scaling can sometimes complicate rather than streamline. Additionally, foundational elements, such as a firm grasp of agile practices and automation, can determine the success of scaling endeavors. In essence, scaling is a tool in the agile toolkit—powerful when used correctly but potentially counterproductive if misapplied. Organizations must reflect on their unique scenarios, understanding both the promises and pitfalls of scaling, to ensure they chart a path that genuinely enhances agility, efficiency, and value delivery.


From Big Data to Better Data: Ensuring Data Quality with Verity

High-quality data is necessary for the success of every data-driven company. It enables everything from reliable business logic to insightful decision-making and robust machine learning modeling. It is now the norm for tech companies to have a well-developed data platform. This makes it easy for engineers to generate, transform, store, and analyze data at the petabyte scale. As such, we have reached a point where the quantity of data is no longer a boundary. Yet this has come at the cost of quality. ... Poor data quality in Hive caused tainted experimentation metrics, inaccurate machine learning features, and flawed executive dashboards. These incidents were hard to troubleshoot, as we had no unified approach to assessing data quality and no centralized repository for results. This delay increased the difficulty and cost of data backfills. The lack of centralization in data quality also made the data discovery process inefficient, making it hard for data scientists and data engineers to identify trustworthy data.


AI vs software outsourcing: An opportunity or a threat?

As AI becomes more widespread, the question is whether programmers write code themselves or have chatbots write it. Customers usually expect quality. If AI can help deliver this quality faster, why not? Look, everyone knows that there is a programming language called Java. There are Apache Commons libraries. You can Google it, but can you do something with it? Can you bring value to the business? This is the point. LLM models are a tool, just like a library or a framework. However, it has other capabilities that need to be mastered and used to bring value. It will be a long time before AI can replace developers because there will always be something that needs to be fixed. Either it's an error in the code or something wrong with the configuration. For example, if a bot has already written code that seems to work, but an error appears. The developer can spend little time writing the code but later spends more time looking for the error. Let's take GitHub Copilot. Programmers note that the acceptance rate of suggestions from Copilot is up to 40%. 


Why all IT talent should be irreplaceable

“Great employee” is easy to type. It’s less easy to define. Here’s a short list to get you started. Scrub it by discussing the question with your leadership team. The habit of success: Some employees seemingly don’t know how to fail. Give them an assignment and they’ll figure out a way to get it done. Competence: As a general rule, it’s better to apologize for an employee’s bad manners than for their inability to do the work. Without competence, employees with a strong success habit can do a lot of damage by, for example, creating kludges instead of sustainable solutions. Followership: Leadership is a prized attribute for employees to have. Prized, that is, if they’re leading in their leader’s direction. Otherwise, if you and they are leading in different directions, all your prized leaders will do is generate conflict and confusion. Followership is what happens when they embrace the direction you’re setting and make it their own. Intellectual honesty: Some employees can be persuaded with evidence and logic. Others trust their guts instead. That’s a physiological error. You want people who digest with their intestines but think with their brains.


Do you need both cloud architects and cloud engineers?

We need a collaborative approach with both disciplines. One cannot function properly without the other. For example, I cannot design multicloud-based systems that define different usages for different cloud services on different clouds. ... Many assume that the engineering tasks are the easiest part of the journey to the cloud. After all, if the cloud architect is good, the configuration should work, and it’s just a matter of using sound AI tools to carry out deployment. Even worse, some companies are working just with engineers and hiring specific skills. The company may pick a cloud brand and hire security, application, data, and AI engineers in that cloud platform. They assume that this specific cloud platform is the correct and optimized platform, which will usually cause trouble. Oh, the solutions may work, but it could cost 10 times more to operate. Not surprisingly, these companies have an underoptimized architecture since they’ve given zero consideration to architecture or the use of cloud architects. AI won’t save you from needing a good architecture and a good set of engineering disciplines. 


What IT needs to know about energy-efficiency directives for data centers

New regulations springing up in various regions will be among the drivers of data center sustainability in the months ahead. There are two main groups of regulations emerging that affect data center operations, according to Jay Dietrich, research director of sustainability at Uptime Institute. One is financial reporting modeled on the Task Force for Climate-related Financial Disclosures (TCFD), which requires reporting on energy consumption and efficiency and associated greenhouse gas (GHG) emissions. The other is the European Energy Efficiency Directive (EED), which requires an energy management plan, an energy audit, and reporting of operational data. In addition, there are voluntary, country-specific standards and siting requirements for data center efficiency and operations in various countries around the world, Dietrich says. A current example of a TCFD-related regulation is the E.U. Corporate Sustainability Reporting Directive (CSRD), with reporting requirements rolling out from large to small enterprises beginning in 2025 and continuing until 2028.


What does leadership in a hybrid world look like?

Firms want their best people to stick around and give more of themselves. Studies have shown that improved employee collaboration and alignment with a common purpose is key to achieving that. But what is the best way to make that happen in the way we now wish to work and live our lives? Some suggest that the emergence of generative AI and new work tools can improve productivity regardless of the workplace setting. But perhaps a different, more human, approach is needed? The profound loosening of relationships that employees have with their firm and one another, requires a similarly fundamental reimagining of the role of the leader itself. Ultimately, this will not come through new technology, systems, processes, or HR policy (however well-crafted), but through the actions and behaviours of credible and engaging people managers. Firms need to re-establish a sense of cohesion and that needs people who are exceptional good at doing just that. Businesses can’t just issue ultimatums or mandates; they need a leadership approach that “coheres” employees to feel less remote from one another and the firm.


Six skills you need to become an AI prompt engineer

Prompt engineering is much more of a collaborative conversation than an exercise in programming. Although LLMs are certainly not sentient, they often communicate in a way that's similar to how you'd communicate with a co-worker or subordinate. When you're defining your problem statements and queries, you will often have to think outside the box. The picture you have in your head may not translate to the internal representation of the AI. You'll need to be able to think about a variety of conversational approaches and different gambits to get the results you want. ... While you might not necessarily be expected to write the full application code, you will provide far more value if you can write some code, test your prompts in the context of the apps you're building, run debug code, and overall be part of the interactive programming process. It will be much easier for a team to move forward if the prompt engineering occurs as an integral part of the process, rather than having to add it in and test it as a completely separate operation.


The Cost Dynamics of Multitenancy

Isolating tenants with infrastructure has a higher initial cost, especially as you discover the right size for tenant workloads. Once you understand the cost for a tenant, it provides a very stable cost per tenant. Any unevenness in the cost profile represents a choice of timing. For example, if you use containers per tenant, you must decide when to commission your next cluster. Software-based multitenancy has an early advantage as it keeps the initial product price low. The marginal economics of onboarding a tenant are very low — almost zero. There comes a point when the initial design can no longer manage the load. The first port of call is vertical scaling — adding more power to the infrastructure to handle the load. This increases the cost per tenant but enables further tenants to be added. Eventually, you run out of vertical scaling options and look to horizontal scaling. This requires more investment as you need to handle load balancing, re-architect stateful interactions and introduce technologies such as shared cache.



Quote for the day:

"When you stop chasing the wrong things you give the right things a chance to catch you." -- Lolly Daskal

Daily Tech Digest - October 03, 2023

How AI can be a ‘multivitamin supplement’ for many industries

It won’t replace humans in the same way that supplements don’t replace a healthy diet. Still, it will strengthen companies’ existing operations and fill in the gaps that are currently making work more burdensome for human laborers. ... It’s exciting to realize that there will soon be professions that we don’t even have names for yet. As the technology ages and matures and governing bodies create the necessary laws and regulations, our current state of uncertainty will transform into an exciting, bright new future of human-tech cooperation. We are already seeing this future take shape. For instance, MarTech companies are testing AI-powered fraud detection to supplement the work that human experts do to monitor traffic quality and transparency. This not only eases the human workload but helps companies save resources while getting better results overall. Similar benefits of human-AI collaboration can be seen in healthcare, with AI that can be trained to assist patients with recovery treatments or perform routine tasks in medical offices or hospitals, freeing nurses and doctors up to focus on patient outcomes. 


Banking on Innovation: How Finance Transforms Technological Growth for Decision Makers

Regulation is a sensitive topic for the financial industry. While the need for a certain degree of oversight is universally accepted, excessive regulation can stifle the very innovation that drives economic growth. On the other hand, too little regulation can open the doors to risk accumulation and financial crises. Striking this balance is one of the most challenging tasks that government leaders face. Policies must be evidence-based, derived from transparent risk- assessment models and economic simulations. Regulatory sandboxes could offer a safe environment for financial institutions to experiment with new services and products under the watchful eye of regulators, thereby fostering innovation while ensuring compliance. ... One of the most potent ways in which PPPs can contribute to revenue management is through asset monetization. Governments often sit on a wealth of underutilized assets, ranging from real estate to utilities. A PPP can unlock the value of these assets by involving private-sector expertise and investment. 


Microsoft Releases Its Own Distro of Java 21

Microsoft’s continuing support for OpenJDK is a strong indicator of how important Java is in the enterprise software space. “And the new features of Java 21 such as lightweight threads are maintaining Java’s relevance in the cloud native age,” said Mike Milinkovich, executive director of the Eclipse Foundation. “Being one of the first vendors to ship Java SE 21 support shows how focused Microsoft is in meeting the needs of Java developers deploying workloads on Azure.” Also, Spring developers will be pleased to know that Spring Boot 3.2 now supports Java 21 features. Many other frameworks and libraries will soon release their JDK 21-supported versions. “Microsoft has some of the best developer tool makers in the world — to have them add Java to the mix makes sense,” said Richard Campbell, founder of Campbell & Associates. “Of course, that happened a couple of years ago, and JDK 21 is just the latest implementation. In the end, Microsoft wants to ensure that Azure is a great place to run Java, so having a team working on Java running in Azure helps to make that true. What does it mean for the ecosystem? More choices for implementations of Java, better Java tooling, and more places to run Java fast and securely.”


Why embracing complexity is the real challenge in software today

The reason we can’t just wish away or “fix” complexity is that every solution — whether it’s a technology or methodology — redistributes complexity in some way. Solutions reorganize problems. When microservices emerged (a software architecture approach where an application or system is composed of many smaller parts), they seemingly solved many of the maintenance and development challenges posed by monolithic architectures (where the application is one single interlocking system). However, in doing so microservices placed new demands on engineering teams; they require greater maturity in terms of practices and processes. This is one of the reasons why we cautioned people against what we call “microservice envy” in a 2018 edition of the Technology Radar, with CTO Rebecca Parsons writing that microservices would never be recommended for adoption on Technology Radar because “not all organizations are microservices-ready.” We noticed there was a tendency to look to adopt microservices simply because it was fashionable. This doesn’t mean the solution is poor or defective. 


Balancing Cost and Resilience: Crafting a Lean IT Business Continuity Strategy

Effective monitoring is the backbone of a resilient infrastructure. The approach should focus on: Filtering out the noise - Monitoring solutions need to ensure that only critical notifications are sent out, preventing information overload and ensuring that the right people are alerted promptly when critical events inevitably happen. Acting quickly and decisively - Time is of the essence during disruptions. IT, DevOps, SIRT, and even PR teams need to be well coordinated for various types of events. From security breaches to data center fires or even just mundane equipment failures, anything that might result in customer or operation disruptions will involve cross-team communications and collaboration. The only way to get better at handling these is to have documentation on what should be done, a clear chain of command, and practice drills. In conclusion, a comprehensive backup and recovery strategy is essential for businesses aiming for uninterrupted operations. While there are many solutions available in the market, it’s crucial to find one that aligns with your business needs. 


How do you solve a problem like payments infrastructure?

Today, banks need to be willing to adopt new technology to change, and this will involve working with a third-party service provider. Another roundtable participant added that as part of this process, it is imperative to utilise validation evaluation to recycle new enhancements. Otherwise, banks will end up with the belief that the improvements that were made are unique, but in fact, competitors will keep pace or even get ahead when it comes to the innovation game or enticing new customers. This banker revealed that they opted to not disconnect from their existing infrastructure, but instead chose a top layer architecture to process payments in a more efficient way. In line with this, the participant added that culture must be considered, because this is what brings together the different components that are needed and ultimately reveals when the time is right to change the systems. Providing background information, this Sibos attendee mentioned that 15 years ago, the bank considered whether it would be more cost effective to map local, regional, or global ISO 20022 messaging into existing architecture or to create a new platform that could work for the next 20 years. 


GenAI: friend or foe in fraud risk management?

Building high-performance fraud detection algorithms today is dependent on real-life customer and transaction data to train and validate the models, which has remained a constraint. GenAI can help with realistic synthetic data creation for model training and validation, scenario and fraud attack simulation to identify vulnerabilities and design controls to mitigate these risks. Customer due diligence (CDD) is a critical function in fraud prevention – be it new client onboarding or new credit approvals (loans, credit cards, increasing credit limits) for existing clients. GenAI can be a great tool to go through piles of KYC documentation and reference them with customer-filled forms and other subscribed data sources of the FI to come up with a CDD summary report. GenAI can also be used to analyse user communications with FIs – such as emails, chats, documents and product and service requests – to extract insights on financial behaviour, sentiment analysis for intentions and potential risks of fraud. Fraud investigations can also leverage GenAI for alert and dispute resolution by accessing different sources of information on the context and providing a summary of the case that will aid in its decisioning.


Weaving Cyber Resilience into the Strategic Fabric of Higher Education Institutions

There is no shortage of steps that institutions can take to bolster their cyber resilience and ensure that, should the worst happen, they’re prepared. A good place to start is by assessing the institution’s current level of resilience and looking for any gaps or obstacles. In many cases, Goerlich says, the key is simplification. For example, adopting a zero-trust security strategy can also improve a college or university’s ability to respond, maintain continuity and bounce back following an adverse event, he says. Another factor complicating resiliency for many institutions is overly complex network environments, particularly in the cloud. As colleges and universities clamor to embrace digital transformation and cloud networking, it’s not uncommon for their environments to grow to a degree that becomes unmanageable. But uncontrolled and unregulated cloud sprawl can have a serious impact on an institution’s resilience. Developing easy-to-follow approaches and processes — along with adopting simplified, automated and easy-to-use technology solutions — can make a significant difference, Goerlich says. 


How to make asynchronous collaboration work for your business

Asynchronous working can bring some benefits that synchronous work can't – most notably speed. “Real-time communication means everyone must be in the same place, or at least the same time zone, in order for work to happen. If workers need to wait for syncs to decide or act on something, it slows down the company as a whole and reduces its ability to compete,” says van der Voort. Asynchronous collaboration allows people to work at their own pace, and does not force them to wait for input from others. Morning people, evening people, midnight oil people, collaborating across geographies, can in some cases deliver higher quality results than forcing everyone to come together for a 10am video call. To get this working well, policies such as having core working hours for each staff member, and having very clear goals and anticipated outcomes for all meetings, can be incredibly useful. “One of the most significant and highly sought-after benefits asynchronous collaboration offers is a dramatic reduction in meetings,” argues Lawyer. “It allows team members to contribute in the least amount of minutes, freeing up time for other work.”


Securing the Evolution of Smart Home Applications

Very few in the cybersecurity community have forgotten one of the most noteworthy incidents, the Mirai Botnet, which took place back in 2017. Attackers behind the botnet infiltrated the site of well-known cybersecurity journalist Brian Krebs. The Distributed Denial of Service (DDoS) attack lasted for days, 77 hours to be exact. It involved 24,000 Mirai-infected Internet-of-Things devices, including personal surveillance cameras. Jumping ahead to this year 2023, in June the Federal Trade Commission (FTC) settled a case with Ring’s owner, Amazon. The online retailing giant agreed to pay the FTC nearly $31 million in penalties to settle recently filed federal lawsuits over privacy violations. The FTC alleged that Ring compromised customer privacy by allowing any employee or contractor to access consumers’ private videos. The FTC also claimed hackers used Ring cameras’ two-way functionality to harass and even physically threaten consumers – including children – if they did not pay a ransom. These types of incidents clearly illustrate how critical it is to secure devices like cameras in a smart home.



Quote for the day:

"Before you are a leader, success is all about growing yourself when you become a leader, success is all about growing others." -- Jack Welch

Daily Tech Digest - October 02, 2023

Want people to embrace transformation? Allow them to own the change

The principles for a co-optable resource are straightforward: for starters, it must be accessible. Accessible means it must be opt-in, no mandates, no obvious carrots or sticks, and it is owned by those opting in. The barriers to entry must be low, and the benefits of using the resource have to be easy to communicate to others. Finally, it must be both impactful—that is, delivering practical value to its users—and scalable. In each of the following examples, a co-optable resource led to widespread uptake of a new idea or technology. The first one shows how a small organization was able to replicate itself globally by sharing the heavy lifting of making an idea scalable—an important lesson for managers who are daunted by introducing new ways of working because they feel the burden is all on them. The other two examples show how it’s possible to get enthusiasts within organizations to scale the use of technology, transform a business model, and change ways of working.


Weed Out Bad Data to Make Better Business Decisions

Using bad data for analytics, AI, and other apps can have catastrophic consequences for any organization. The worst-case scenario is making poor business decisions with that data – whether it’s investments, product changes, or hiring moves. Ignoring and not removing bad data results in misleading insights and misguided choices. It’s like blindly following a GPS without verifying its accuracy or knowing its end goal. You could potentially drive yourself into the ocean. It also has a broader chilling effect on a company. When bad data leads to skewed or inaccurate insights, employees lose trust in the data and systems more broadly. As a result, they stop relying on the data to make decisions altogether and instead devolve to making decisions based on gut feeling. At a bare minimum, bad data should be weeded out as often as you use it to make decisions. Ideally, though, it should happen upon the ingestion of the data. Constantly removing bad data as soon as it enters the system is the only way to reliably avoid polluting the clean data source.


California’s Delete Act: What, CIOs, CDOs, Businesses Need to Know

The bill says consumers can delete data by using a website that will be hosted by the California Privacy Protection Agency, which has a 2026 deadline to create the website. In 2026, data brokers registered with the state must process delete requests once a month and undergo third-party audits every three years starting in 2028. Brokers who don’t comply will face daily fines. California’s law is not the first state law to target data brokers. Vermont, Texas, and Oregon all have laws creating broker registries. Vermont’s law has been in effect since 2019. California’s Data Broker law defines a data broker as “a business that knowingly collects and sells to third parties the personal information of a consumer with whom the business does not have a direct relationship.” While there is a federal data privacy bill, the American Data Privacy Protection Act (ADPPA), the proposal is currently in US Congress limbo and chances for passage are unclear. ADPPA would instruct the Federal Trade Commission (FTC) to create a national registry of data brokers and create a “do not collect” mechanism for individuals to opt out of personal data collection.


Global events fuel DDoS attack campaigns

NETSCOUT’s insights into the threat landscape come from its ATLAS sensor network built over decades of working with hundreds of Internet Service Providers globally, gleaning trends from an average of 424 Tbps of internet peering traffic, an increase of 5.7% over 2022. The company has observed nearly 500% growth in HTTP/S application layer attacks since 2019 and 17% growth in DNS reflection/amplification volumes during the first half of 2023. “While world events and 5G network expansion have driven an increase in DDoS attacks, adversaries continue to evolve their approach to be more dynamic by taking advantage of bespoke infrastructure such as bulletproof hosts or proxy networks to launch attacks,” stated Richard Hummel, senior threat intelligence lead, NETSCOUT. “The lifecycle of DDoS attack vectors reveals the persistence of adversaries to find and weaponize new methods of attack, while DNS water torture and carpet-bombing attacks have become more prevalent.”


Multibillion-dollar cybersecurity training market fails to fix the supply-demand imbalance

The good news is that retention increased, with a 6% drop in the number of respondents reporting retention issues compared to the previous year. But this improvement is more likely tied to economic uncertainty rather than work conditions having improved. The main reasons for employees departing included recruitment by other companies (58%). The second highest response, poor financial incentives (e.g., salaries or bonuses), is likely the main driver, ISACA found. Those seeking better financial compensation increased by 6% from last year to 54%. While work stress levels dropped by two percentage points from 2022, it remains a contributing factor at 43%, ranking fourth on the list. Other notable reasons included limited remote work possibilities (increased by four percentage points from 2022) and poor work culture/environment, both potentially driven by return-to-work mandates. "Uncertainty of any kind appears to be driving fewer job changes, and while vacancies persist, the survey results indicate that enterprises appear to be tightening budgets and compensation aids ahead of a potential recession," read the report.


Prompt Engineering in Software Automation

While these problems can’t be ignored, there is still a lot of justifiable excitement about how these programs can help democratize software development by supporting technical and non-technical teams alike. Perhaps the most impressive thing to consider is that tools like ChatGPT can produce functional code very quickly. With the right prompt, engineers can reduce the time it takes to program certain types of code, ensuring a swifter software development life cycle. At the end of 2022, the popular programming hub Stack Overflow banned AI-generated answers on its forum. They cited the high error rate and inaccuracies associated with the application. However, the technology is in a nascent stage; furthermore, the dissatisfaction with AI-generated output owes as much to poor prompt engineering as it does to the technology itself. Despite the misgivings over the tech, a recent piece by McKinsey highlights the impact that prompt engineering is already having in the world of programming. The consulting firm’s 


Hackers Impersonate Meta Recruiter to Target Aerospace Firm

The attack is part of an ongoing campaign tracked as "Operation DreamJob," in which fake recruiters reach out through LinkedIn. Attackers convince victims to self-compromise their systems by employing different strategies such as luring the target to execute a malicious PDF viewer to see the full contents of a job offer. Or, they encourage the victim to connect with a Trojanized SSL/VPN client. "The most worrying aspect of the attack is the new type of payload, LightlessCan, a complex and possibly evolving tool that exhibits a high level of sophistication in its design and operation, representing a significant advancement in malicious capabilities compared to its predecessor, BlindingCan," researchers said. Eset says is observed victims receiving two malicious executables, Quiz1.exe and Quiz2.exe, which were delivered via .iso images hosted on a third-party cloud storage platform. "The first challenge is a very basic project that displays the text 'Hello, World!'" researchers said. "The second prints a Fibonacci sequence up to the largest element smaller than the number entered as input."


Technology is Crack and We are the Dealers

What is actually going on though is not really sinister, it is just stupid. For years most technology did not really impact lives outside of military, reactors, planes, infrastructure and the like… then medicine, electrical grids, and finances joined the group. And so forth. But most technology was just corporate enablement. No one was going to die if the order management system went down for an hour. Maybe get fired but not die. Thus we chose to use standards and review (governance) as our primary mechanism for quality decisions. And even these were flaky at best and pretty easy to get around (please like I can’t game a governance review board? hahahaha). The people reviewing had their checklists and the delivery folks knew how to make them happy enough. Or just go to the executive sponsor who goes to the executives and gets a ‘pass’. Oh well, it’s just a bit of technical debt! The future is coming to humanity. That much is certain. But at what rate? What is acceptable loss? How will society get a handle on run-away technology? And which organizations will survive? 


The dark arts of digital transformation — and how to master them

“If you’re in a leadership role in Engineering, you aren’t going to succeed unless you have a strong ally in Product,” says Etkin. “Developers sometimes have this idea that management isn’t necessary, or they have disdain for the nontechnical side of things. That’s a terrible idea that will get you absolutely nowhere.” Etkin, an early employee at Atlassian who was the original architect of Jira, admits that he wasn’t always good at building alliances with his peers. He had to figure out how to get on the same page with people who often had very different ideas about how to proceed. That meant asking a lot of questions and listening to the answers. ... A key thing to remember is that the dark forces you’re attempting to subdue may not be the individuals opposing you, but the systems in which they themselves are trapped. Organizations that have found success operating in a certain way may see little reason to shake things up. Even when the changes are necessary, such as in the case of increased competition from disruptive new entrants or the emergence of transformative technologies, the effort required to overcome internal inertia could exhaust all your magic powers.


Regulations Push Firms to Boost AI, ML Spend

Unlike some industries, though, financial services are highly regulated, given the industry’s stature as the modern economy’s backbone. “The industry as a whole must be cautious about adopting new technologies given the myriad of rules and regulations at play,” cautions Joe Robinson, CEO, Hummingbird. “Financial institutions can plan to leverage the opportunities that AI presents but must do so carefully.” He says by using explainable algorithms, auditable decision-making processes, and/or human-in-the-loop reviews, they can take advantage of the potential of AI while ensuring that regulatory obligations are met. “As with many new technologies, it's best to start small, observe outcomes, and scale up thoughtfully and pragmatically,” he says. Cullen adds it’s critical to ensure the needed talent infrastructure is in place. “Determine where you should hire and where you may need to augment, especially in relation to the evolving regulatory landscape,” she says.



Quote for the day:

”Taking a step back can often be the quickest way forward.” — Tim Fargo

Daily Tech Digest - October 01, 2023

The future of work is human-AI synergy

AI and humans can work in sync by capitalising on their respective strengths. AI's ability to automate routine tasks liberates human workers to focus on more complex and nuanced responsibilities, where their human touch is indispensable. This dynamic significantly amplifies productivity and allows employees to dedicate their time to strategic thinking and fostering innovation. AI's application in Big Data Analytics equips human workers with invaluable insights, enabling them to make quicker, more informed decisions with heightened precision. For instance, financial institutions employ AI analytics to rapidly evaluate loan applications, while healthcare professionals use AI algorithms to swiftly diagnose serious illnesses from patient data. However, it's crucial to emphasise that AI serves as a valuable tool rather than a replacement for human workers. The efficiency and productivity gains result from the synergy between human intelligence and AI capabilities.


10 Strategies for Simplified Data Management

Centralization means creating a unified, accessible, and authoritative store for all of your organizational data. Users and processes can then leverage and manage otherwise distinct data in a convenient, coherent fashion. The two main approaches here are data lakes and data warehouses. A data lake is a large repository of different kinds of data - all stored in their original format. This provides a valuable resource, as we can apply any kind of transformations and aggregation we need for analysis. A data warehouse differs from a data lake in the sense that it is stored in a format and structure that’s defined for a specific purpose. This is useful if we need to carry out similar analytical operations on a large scale. ... As we said earlier, an enterprise data model is a detailed account of all of the data assets that are involved in core business processes - along with where each of these is sourced from, what they’re used for, and how they relate to each other. This is effectively a data-centric representation of how your business works. In turn, an effective data model brings along several important benefits. 


Data Quality Assessment: Measuring Success

A Data Quality assessment will move along more efficiently and provide better results if a list of concerns and goals is created before the assessment. When creating this list, be aware of the organization’s long-term goals, while listing short-term goals. For example, the long-term goal of making the business more efficient can be broken down into smaller goals, such as fixing the system so the right people get the right bills, and that all the clients’ addresses are correct, etc. This list can also be presented to a board of directors as a rationale for initiating and paying for Data Quality assessment software or hiring a contractor to perform the assessment. The basic steps for creating the list are presented below.Start by making a list of Data Quality problems that have occurred over the last year. Spend a week or two observing the flow of data and determine what looks questionable, and why. Share your observations with other managers and staff, get feedback, and adjust the results using the feedback. 


Test Architecture: Creating an Architecture for Automated Tests

The test architecture is important, especially when you are dealing with a complex project or expecting the project to grow in the near future. The test architecture helps to reduce the risks and eliminate the assumptions before delivery. As you are aware, anything you do randomly may not help in a better outcome. The test architecture streamlines the entire process of testing. Unlike other testing activities, it’s not focused on a single testing activity rather it is focused on the entire testing and the testing team aims to deliver a high-quality product. ... The test architect works with multiple teams such as development, DevOps, testing, and business/product team. So the test architect is responsible for communication with stakeholders. If there are any challenges from the development team he should be able to work with them and get them resolved. The complexity of the test architecture for automation depends on the tool you choose. Because some tools require creating the framework, some come with a framework ready. Not all the tools require coding, so the activities that are involved in defining the coding standards and setup will be reduced. 


7 Cybersecurity Questions That Can Transform Your Business

Anyone who has spent any time thinking about cybersecurity knows how multifaceted and complicated our digital supply chains are today. That means we need to empower people who are working directly with the different touch points in the supply chain and elevate their cybersecurity thinking. They need all the information and resources available to ensure they only push secure software to customers. ... You may have a long list of audits and other compliance procedures in progress currently. This is where I ask you to remember that the point of a canvas like this is that it is one page! While that may not give you all the room to include every initiative, that may be a good thing. Instead of starting new small-scale initiatives, consider, for instance, adopting or enhancing a DevSecOps approach that could transform your security efforts. ... When we talk about costs here, we mean actual costs. This includes external consultant fees, CISO office salaries, MDR subscriptions, security training and platform subscriptions. When confronted with these numbers, we can make decisions that aren’t only guided by whims or immediate needs.


Why Cloud Native Expertise Is so Hard to Hire for, and What to Do Instead

Fortunately, there are alternatives for organizations looking to develop their cloud native expertise. One of the most popular options is to work with a third-party provider that specializes in providing cloud native services — so you don’t have to. This is a core component of what entails “ZeroOps,” or rather, the notion of freeing your own employees to take their time back, and letting someone else do the time-consuming, bothersome stuff. Working with a third-party provider can provide organizations with high levels of expertise and resources while allowing your team to focus on their core business — innovating, creating, and making a measurable impact. This can result in significant cost savings and increased efficiency, as the provider takes on the responsibility of managing complex cloud native solutions. Many providers can offer comprehensive services, ranging from architecture to software engineering and deployment, and can tailor their services to an organization’s unique requirements — of which we know there are many.


The CISO Carousel and Its Effect on Enterprise Cybersecurity

“There is still a prevalent perception that CISOs are viewed as scapegoats in serious breach events,” adds George Jones, CISO at Critical Start. “This is based on a general lack of understanding, high expectations, and accountability associated with the role. When a breach occurs, it’s easy to point the finger at the person responsible for cybersecurity.” It’s the effect, says Yu, of “accountability without authority”. Making the CISO a scapegoat is a common but not blanket response to cybersecurity incidents. Agnidipta Sarakar, VP and CISO advisory at ColorTokens, points out, “Organizations who are mature tend not to blame the CISO unless the security program is actually not good enough.” But less mature organizations with weaker programs or negligent security oversight will readily activate the scapegoat effect. ... Globally, there are many companies where cybersecurity is both prioritized and supported, but these tend to be among the larger and more mature organizations. There remains a large underswell of newer and smaller companies where growth is often prioritized over security.


Closing the skills gap in the AI era: A global imperative

To tackle this reskilling challenge on a large scale, we require a combined effort from the government, education, and private sector. This can be achieved through the following ways: Make learning achievable: Instead of diving into the deeply technical aspects of AI, companies can begin by introducing the workforce to tools that require no-code or low-code experience Further, citizen development programs can be implemented. These programs encourage employees to be innovative problem solvers and foster a sense of ownership as they witness the direct impact of their work on business outcomes using no-code/low-code tools. These programs allow them to savour initial automation successes almost immediately and to envision greater possibilities for bots to help them in the future. Take advantage of existing partnerships: Companies should leverage the knowledge of their existing technology partners to quickly roll out skilling programs. The National Health Service in the UK, for example, was able to offer its 1.7 million employees automation training via the help of its technology partner.


Could APIs undermine Zero Trust?

APIs come in various shapes and flavours. As well as being internal or public facing, they might interface in numerous ways, from a single API providing access to a service mechanism, to aggregated APIs that then use another as the point of entry, to APIs that act as the go-between between various non-compatible applications, or partner/third party APIs. They are also problematic to monitor and secure using traditional mechanisms. Segmentation and deep inspection technology at the layer 7 network level can miss APIs completely, resulting in those shadow APIs, while application level 4 protection methods such as web application firewalls (WAFs) which use signature-based threat detection will miss the kind of abuse that typically leads to API compromise. Often, APIs are not’ hacked’ as such, but their functionality is used against them in business logic abuse attacks and so it’s the behaviour of the API request and resulting traffic that needs to be observed. Yet it’s clear that APIs must be included in ZTA. 


Transforming Decision-Making Processes

GenAI has quickly become a part of everyday conversations from the boardroom to the kitchen table. One specific topic of interest is the role genAI can play in enhancing and improving an organization’s decision-making paradigm. Organizations should look for AI engines that combine the power of artificial intelligence, machine learning, and generative AI to further advance the democratization of analytics. This can reduce the time required to derive insights from data. With AI and cloud-native analytics automation, the power and scale of better decision-making is at everyone’s fingertips. While it is still the early days for genAI, we see this newer capability accelerating the path for organizations to become more insights driven in their decision-making. Natural language processing translates insights into business language that can be shared broadly and leveraged by all. GenAI and large language models (LLMs) eliminate tedious tasks, leverage best practices from millions of workflows in production, automatically document workflows, and free up time for humans to focus on more strategic challenges.



Quote for the day:

"People often say that motivation doesn't last. Well, neither does bathing - that's why we recommend it daily." -- Zig Ziglar