Daily Tech Digest - November 02, 2024

Cisco takes aim at developing quantum data center

On top of the quantum network fabric effort, Cisco is developing a software package that includes the best way for entanglement, distribution effort, protocol, and routing algorithms, which the company is building in a protocol stack and compiler, called Quantum Orchestra. “We are developing a network-aware quantum orchestrator, which is this general framework that takes quantum jobs in terms of quantum circuits as an input, as well as the network topology, which also includes how and where the different quantum devices are distributed inside the network,” said Hassan Shapourian, Technical Leader, Cisco Outshift. “The orchestrator will let us modify a circuit for better distributability. Also, we’re going to decide which logical [quantum variational circuit] QVC to assign to which quantum device and how it will communicate with which device inside a rack.” “After that we need to schedule a set of switch configurations to enable end-to-end entanglement generations [to ensure actual connectivity]. And that involves routing as well as resource management, because, we’re going to share resources, and eventually the goal is to minimize the execution time or minimize the switching events, and the output would be a set of instructions to the switches,” Shapourian said.


How CIOs Can Fix Data Governance For Generative AI

When you look at it from a consumption standpoint, the enrichment of AI happens as you start increasing the canvas of data it can pick up, because it learns more. That means it needs very clean information. It needs [to be] more accurate, because you push in something rough, it’s going to be all trash. Traditional AI ensured that we have started cleaning the data, and metadata told us if there is more data available. AI has started pushing people to create more metadata, classification, cleaner data, reduce duplicates, ensure that there is a synergy between the sets of the data, and they’re not redundant. It’s cleaner, it’s more current, it’s real-time. Gen AI has gone a step forward. If you want to contextually make it rich, you want to pull in more RAGs into these kinds of solutions, you need to know exactly where the data sits today. You need to know exactly what is in the data to create a RAG pipeline, which is clean enough for it to generate very accurate answers. Consumption is driving behavior. In multiple ways, it is actually driving organizations to start thinking about categorization, access controls, governance. [An AI platform] also needs to know the history of the data. All these things have started happening now to do this because this is very complex.


Here’s the paper no one read before declaring the demise of modern cryptography

With no original paper to reference, many news outlets searched the Chinese Journal of Computers for similar research and came up with this paper. It wasn’t published in September, as the news article reported, but it was written by the same researchers and referenced the “D-Wave Advantage”—a type of quantum computer sold by Canada-based D-Wave Quantum Systems—in the title. Some of the follow-on articles bought the misinformation hook, line, and sinker, repeating incorrectly that the fall of RSA was upon us. People got that idea because the May paper claimed to have used a D-Wave system to factor a 50-bit RSA integer. Other publications correctly debunked the claims in the South China Morning Post but mistakenly cited the May paper and noted the inconsistencies between what it claimed and what the news outlet reported. ... It reports using a D-Wave-enabled quantum annealer to find “integral distinguishers up to 9-rounds” in the encryption algorithms known as PRESENT, GIFT-64, and RECTANGLE. All three are symmetric encryption algorithms built on a SPN—short for substitution-permutation network structure.


AI Has Created a Paradox in Data Cleansing and Management

When asked about the practices required to maintain a cleansed data set, Perkins-Munn states that in that state, it is critical to think about enhancing data cleaning and quality management. Delving further, she states that there are many ways to maintain it over time and discusses a few that include AI algorithms revolving around automated data profiling and anomaly detection. Particularly in the case of unsupervised learning models, AI algorithms automatically profile data sets and detect anomalies or outliers. Continuous data monitoring is one ongoing way to keep data clean. She also mentions intelligent data matching and deduplication, wherein machine learning algorithms improve the accuracy and efficiency of data matching and duplication processes. Apart from those, there are fuzzy matching algorithms that can identify and merge duplicate records even with minimal variations or errors. Moving forward, Perkins-Munn states that for effective data management, organizations must prioritize where to start with data cleansing, and there is no one-method-fits-all approach to it. She advises to focus on cleaning the data that directly impacts the most critical business process or decision, thus ensuring quick, tangible value.


A brief summary of language model finetuning

For language models, there are two primary goals that a practitioner will have when performing fine tuning: Knowledge injection: Teach the model how to leverage new sources of knowledge (not present during pretraining) when solving problems. Alignment (or style/format specification): Modify the way in which the language model surfaces its existing knowledge base; e.g., abide by a certain answer format, use a new style/tone of voice, avoid outputting incorrect information, and more. Given this information, we might wonder: Which fine-tuning techniques should we use to accomplish either (or both) of these goals? To answer this question, we need to take a much deeper look at recent research on the topic of fine tuning. ... We don’t need tons of data to learn the style or format of output, only to learn new knowledge. When performing fine tuning, it’s very important that we know which goal—either alignment or knowledge injection—that we are aiming for. Then, we should put benchmarks in place that allow us to accurately and comprehensively assess whether that goal was accomplished or not. Imitation models failed to do this, which led to a bunch of misleading claims/results!  
 

Bridging Tech and Policy: Insights on Privacy and AI from IndiaFOSS 2024

Global communication systems are predominantly managed and governed by major technology corporations, often referred to as Big Tech. These organizations exert significant influence over how information flows across the world, yet they lack a nuanced understanding of the socio-political dynamics in the Global South. Pratik Sinha, co-founder at Alt News, spoke about how this gap in understanding can have severe consequences, particularly when it comes to issues such as misinformation, hate speech, and the spread of harmful content. ... The FOSS community is uniquely positioned to address these challenges by collaboratively developing communication systems tailored to the specific needs of various regions. Pratik suggested that by leveraging open-source principles, the FOSS community can create platforms (such as Mastodon) that empower users, enhance local governance, and foster a culture of shared responsibility in content moderation. In doing so, they can provide viable alternatives to Big Tech, ensuring that communication systems serve the diverse needs of communities rather than being controlled by a handful of corporations with a limited understanding of local complexities.


Revealing causal links in complex systems: New algorithm reveals hidden influences

In their new approach, the engineers took a page from information theory—the science of how messages are communicated through a network, based on a theory formulated by the late MIT professor emeritus Claude Shannon. The team developed an algorithm to evaluate any complex system of variables as a messaging network. "We treat the system as a network, and variables transfer information to each other in a way that can be measured," Lozano-DurĂ¡n explains. "If one variable is sending messages to another, that implies it must have some influence. That's the idea of using information propagation to measure causality." The new algorithm evaluates multiple variables simultaneously, rather than taking on one pair of variables at a time, as other methods do. The algorithm defines information as the likelihood that a change in one variable will also see a change in another. This likelihood—and therefore, the information that is exchanged between variables—can get stronger or weaker as the algorithm evaluates more data of the system over time. In the end, the method generates a map of causality that shows which variables in the network are strongly linked. 


Proactive Preparation: Learning From Crowdstrike Chaos

You can’t plan for every scenario. However, having contingency plans can significantly minimise disruption if worse case scenarios occur. Clear guidance, such as knowing who to speak to about the situation and when during outages, can help financial organisations quickly identify faults in their supply chains and restore services. ... Contractual obligations with software suppliers provide an added layer of protection if issues arise. These ensure that there’s a legally binding agreement in place to ensure suppliers handle the issue effectively. Escrow agreements are also key. They protect the critical source code behind applications by keeping a current copy in escrow and can help organisations manage risk if a supplier can no longer provide software or updates. ... supply chains are complex. Software providers also rely on their own suppliers, creating an interconnected web of dependencies. Organisations in the sector should understand their suppliers’ contingency plans to handle disruptions in their wider supply chain. Knowing these plans provides peace of mind that suppliers are also prepared for disruptions and have effective steps in place to minimise any impact.


AI Drives Major Gains for Big 3 Cloud Giants

"Over the last four quarters, the market has grown by almost $16 billion, while over the previous four quarters the respective figure was $10 billion," John Dinsdale, chief analyst at Synergy Research Group, wrote in a statement. "Given the already massive size of the market, we are seeing an impressive surge in growth." ... The Azure OpenAI Service emerged as a particular bright spot, with usage more than doubling over the past six months. AI-based cloud services overall are helping Microsoft's cloud business. ... According to Pichai, Google Cloud's success is focused around five strategic areas. First, its AI infrastructure demonstrated leading performance through advances in storage, compute, and software. Second, the enterprise AI platform, Vertex, showed remarkable growth, with Gemini API calls increasing nearly 14 times over a six-month period. ... Looking ahead, AWS plans increased capital expenditure to support AI growth. "It is a really unusually large, maybe once-in-a-lifetime type of opportunity," Jassy said about the potential of generative AI. "I think our customers, the business, and our shareholders will feel good about this long term that we're aggressively pursuing it."


GreyNoise: AI’s Central Role in Detecting Security Flaws in IoT Devices

GreyNoise’s Sift is powered by large language models (LLMs) that are trained on a massive amount of internet traffic – including which targets targeting IoT devices – that can identify anomalies in the traffic that traditional system could miss, they wrote. They said Sift can spot new anomalies and threats that haven’t been identified or don’t fit the known signatures of known threats. The honeypot analyzes real-time traffic and uses the vendor’s proprietary datasets and then runs the data through AI systems to separate routine internet activity from possible threats, which whittles down what human researchers need to focus on and delivers faster and more accurate results. ... The discovery of the vulnerabilities highlights the larger security issues for an IoT environment that number 18 billion devices worldwide this year and could grow to 32.1 billion by 2030. “Industrial and critical infrastructure sectors rely on these devices for operational efficiency and real-time monitoring,” the GreyNoise researchers wrote. “However, the sheer volume of data generated makes it challenging for traditional tools to discern genuine threats from routine network traffic, leaving systems vulnerable to sophisticated attacks.”



Quote for the day:

"If you're not confused, you're not paying attention." -- Tom Peters

No comments:

Post a Comment