Daily Tech Digest - March 31, 2024

Experts Concerned by Signs of AI Bubble

"Capital continues to pour into the AI sector with very little attention being paid to company fundamentals," he wrote, "in a sure sign that when the music stops there will not be many chairs available." It's been a turbulent week for AI companies, highlighting what sometimes seems like unending investor appetite for new AI ventures. Case in point is Cohere, one of the many startups focusing on generative AI, which is reportedly in late-stage discussions that would value the venture at a whopping $5 billion. Then there's Microsoft, which has already made a $13 billion bet on OpenAI, as well as hiring most of the staff from AI startup Inflection AI earlier this month. The highly unusual deal — or "non-acquisition" — raised red flags among investors, leading to questions as to why Microsoft didn't simply buy the company. According to Windsor, companies "are rushing into anything that can be remotely associated with AI." Ominously, the analyst wasn't afraid to draw direct lines between the ongoing AI hype and previous failed hype cycles. "This is precisely what happened with the Internet in 1999, autonomous driving in 2017 and now generative AI in 2024," he wrote.


A New Way to Let AI Chatbots Converse All day Without Crashing?

Typically, an AI chatbot writes new text based on text it has just seen, so it stores recent tokens in memory, called a KV Cache, to use later. The attention mechanism builds a grid that includes all tokens in the cache, an “attention map” that maps out how strongly each token, or word, relates to each other token. Understanding these relationships is one feature that enables large language models to generate human-like text. But when the cache gets very large, the attention map can become even more massive, which slows down computation. Also, if encoding content requires more tokens than the cache can hold, the model’s performance drops. For instance, one popular model can store 4,096 tokens, yet there are about 10,000 tokens in an academic paper. To get around these problems, researchers employ a “sliding cache” that bumps out the oldest tokens to add new tokens. However, the model’s performance often plummets as soon as that first token is evicted, rapidly reducing the quality of newly generated words. In this new paper, researchers realized that if they keep the first token in the sliding cache, the model will maintain its performance even when the cache size is exceeded.


AI and compliance: The future of adverse media screening in FinTech

Unlike static databases, AI systems continuously learn and adapt from the data they process. This means that they become increasingly effective over time, able to discern patterns and flag potential risks with greater accuracy. This evolving intelligence is crucial for keeping pace with the sophisticated techniques employed by individuals or entities trying to circumvent financial regulations. Furthermore, the implementation of AI in adverse media screening fosters a more robust compliance framework. It empowers FinTech companies to preemptively address potential regulatory challenges by providing them with actionable insights. This proactive approach to compliance not only safeguards the institution but also ensures the integrity of the financial system at large. Despite the promising benefits, the integration of AI into adverse media screening is not without challenges. Concerns regarding data privacy, the potential for bias in AI algorithms, and the need for transparent methodologies are at the forefront of discussions among regulators and companies alike. 


How to Get Tech-Debt on the Roadmap

Addressing technical debt is costly, emphasizing the need to view engineering efforts beyond the individual contributor level. Companies often use the revenue per employee metric as a gauge of the value each employee contributes towards the company's success. Quick calculations suggest that engineers, who typically constitute about 30-35% of a company's workforce, are expected to generate approximately one million dollars in revenue each through their efforts. ... Identifying technical debt is complex. It encompasses customer-facing features, such as new functionalities and bug fixes, as well as behind-the-scenes work like toolchains, testing, and compliance, which become apparent only when issues arise. Additionally, operational aspects like CI/CD processes, training, and incident response are crucial, non-code components of system management. ... Service Level Objectives (SLOs) stand out as a preferred tool for connecting technical metrics with business value, primarily because they encapsulate user experience metrics, offering a concrete way to demonstrate the impact of technical decisions on business outcomes.


What are the Essential Skills for Cyber Security Professionals in 2024?

It perhaps goes without saying, but technical proficiency is key. It is essential to understand how networks function, and have the ability to secure them. This should include knowledge of firewalls, intrusion detection and prevention systems, VPNs and more. Coding and scripting is also crucial, with proficiency in languages such as Python, Java, or C++ invaluable for cybersecurity professionals. ... Coding skills can enable professionals to analyse, identify, and fix vulnerabilities in software and systems, essential when carrying out effective audits of security practices. It will also be needed for evaluation of new technology being integrated into the business, to implement controls to diminish any risk in its operation. ... There’s no escaping the fact that data in general is the lifeblood of modern business. As a result, every cyber security professional will benefit from learning data analysis skills. This does not mean becoming a data scientist, but upskilling in areas such as statistics that can have a profound impact on your job. At the very least you need to be able to understand what the data is telling you. Otherwise, you’re simply following what other people – usually in the data team – tell you.


Removing the hefty price tag: cloud storage without the climate cost

Firstly, businesses should consider location. This means picking a cloud storage provider that’s close to a power facility. This is because distance matters. If electricity travels a long way between generation and use, a percentage is lost. In addition, data centers located in underwater environments or cooler climates can reduce the energy required for cooling. Next, businesses should ask green providers about what they’re doing to minimize their environmental impact. For example, powering their operations with solar, wind, or biofuels reduces reliance on fossil fuels and so lowers GHG emissions. Some facilities will house large battery banks to store renewable energy and ensure a continuous, eco-friendly power supply. Last but certainly not least, technology offers a powerful avenue for enhancing the energy efficiency of cloud storage. Some providers have been investing in algorithms, software, and hardware designed to optimize energy use. For instance, introducing AI and machine learning algorithms or frequency scaling can drastically improve how data centers manage power consumption and cooling.


RESTful APIs: Essential Concepts for Developers

One of these principles is called statelessness, which emphasizes that each request from a client to the server must contain all the information necessary to understand and process the request. This includes the endpoint URI, the HTTP method, any headers, query parameters, authentication tokens, and session information. By eliminating the need for the server to store session state between requests, statelessness enhances scalability and reliability, making RESTful APIs ideal for distributed systems. Another core principle of REST architecture is the uniform interface, promoting simplicity and consistency in API design. This includes several key concepts, such as resource identification through URIs (endpoints), the use of standard HTTP methods like GET, POST, PUT, and DELETE for CRUD operations, and the adoption of standard media types for representing resources. ... RESTful API integration allows developers to leverage the functionalities of external APIs to enhance the capabilities of their applications. This includes accessing external data sources, performing specific operations, or integrating with third-party services.


The dawn of eco-friendly systems development

First, although we are good at writing software, we struggle to create software that optimally utilizes hardware resources. This leads to inefficiencies in power consumption. In the era of cloud computing, we view hardware resources as a massive pool of computing goodness readily available for the taking. No need to think about efficiency or optimization. Second, there has been no accountability for power consumption. Developers and devops engineers do not have access to metrics demonstrating poorly engineered software’s impact on hardware power consumption. In the cloud, this lack of insight is usually worse. Hardware inefficiency costs are hidden since the hardware does not need to be physically purchased—it’s on demand. Cloud finops may change this situation, but it hasn’t yet. Finally, we don’t train developers to write efficient code. A nonoptimized versus a highly optimized application could be 500% more efficient in power consumption metrics. I have watched this situation deteriorate over time. We had to write efficient code back in the day because the cost and availability of processors, storage, and memory were prohibitive and limited. 


10 Essential Measures To Secure Access Credentials

Eliminating passwords should be every organization’s goal, but if that’s not a near-term possibility, implement advanced password policies that exceed traditional standards. Encourage the use of passphrases with a minimum length of 16 characters, incorporating complexity through variations in character types, as well as password changes every 90 days. Additionally, advocate for the use of password managers and passkeys, where possible, and educate end users on the dangers of reusing passwords across different services. ... Ensure the encryption of credentials using advanced cryptographic algorithms and manage encryption keys securely with hardware security modules or cloud-based key management services. Implement zero-trust architectures to safeguard encrypted data further. ... Implement least privilege access controls, which give users only the minimum level of access needed to do their jobs, and regularly audit permissions. Use just-in-time (JIT) and just-enough-access (JEA) principles to dynamically grant access as needed, reducing the attack surface by limiting excessive permissions.


What car shoppers need to know about protecting their data

You can ask representatives at the dealership about a carmaker’s privacy policies and if you have the ability to opt-in or opt-out of data collection, data aggregation and data monetization — or the selling of your data to third-party vendors, said Payton. Additionally, ask if you can be anonymized and not have the data aggregated under your name and your vehicle’s unique identifying number, she said. People at the dealership “might even point you towards talking to the service manager, who often has to deal with any repairs and any follow up and technical components,” said Drury. ... These days, many newer vehicles essentially have an onboard computer. If you don’t want to be tracked or have vehicle data collected and shared, you might find instructions in your owner’s manual on how to wipe out your personalized data and information from the onboard computer, said Payton. “That can be a great way if you are already in a car and you love the car, but you don’t like the data tracking,” she said. While you may not know if the data was already collected and sent out to third parties, you could do this on a periodic basis, she said.



Quote for the day:

"Thinking should become your capital asset, no matter whatever ups and downs you come across in your life." -- Dr. APJ Kalam

No comments:

Post a Comment