Leaders should borrow an important concept from the project management world: Go slow to go fast. There is often a rush to dive in at the beginning of a project, to start getting things done quickly and to feel a sense of accomplishment. This desire backfires when stakeholders are overlooked, plans are not validated, and critical conversations are ignored. Instead, project managers are advised to go slow — to do the work needed up front to develop momentum and gain speed later in the project. The same idea helps reframe notions about how to lead organizational change successfully. Instead of doing the conceptual work quickly and alone, leaders must slow down the initial planning stages, resist the temptation and endorphin rush of being a “heroic” leader solving the problem, and engage people in frank conversations about the trade-offs involved in change. This does not have to take long — even just a few days or weeks. The key is to build the capacity to think together and to get underlying assumptions out in the open. Leaders must do more than just get the conversation started. They also need to keep it going, often in the face of significant challenges.
Machine learning and artificial intelligence for cybersecurity. It's like two ways we are tackling the problems. For example, machine learning has good results in machine vision, computer vision, or you can look at the games. For example, chess was built that was using machine learning and artificial intelligence. As a result the chess game, played by computer, beat the smartest human a couple of years ago. And it has been a very promising application of AI and machine learning. So at the same time, you can look at how machine learning algorithms are compromised. If you recall, there was a Tesla car speeding. And I think it slipped by at 55 miles on the road [with a speed limit] of 35 miles per hour, just because of a smart piece of tape [on a sign]. In that case we are trying to use the advantage or the benefits that AI and machine learning offer for securing systems that we are envisioning in the years to come. But again, at the same time, can we use the other way around too, right. That's why my research is focusing on AI and machine learning for cybersecurity. At the same time, cybersecurity for AI because we want to secure the system that is working for a greater good. Another example that I could explain is if you are using, let's say a machine learning algorithm to filter out the applicants from the application pool to hire somebody.
Former CIO Dave Kieffer says that “cloud ERP can't be customized, but they can be extended. CRM's can provide full platforms. Extension, not customization, should be the goal no matter the platform.” However, Sacolick suggests when “CIO can seamlessly, integrate data and workflow with ERP and CRM, then customization will be less needed. It can become an architecture decision that provides a lot more business flexibility. Low Code and No Code apps, integrations are one approach.” Deb Gildersleeve agrees and says, “a lot of these legacy systems can’t be customized, and for those that can be, most organizations don’t want to spend the time, money and resources needed to customize them. That’s where low code can come in to complement these systems and work within your existing tech stack.” Francis, meanwhile, suggests, “there will likely always be need for an amount of high coding. Sacolick says, “I call highcoding, DIYCoding. But the real challenge is getting app developers on board with using low code /no code where it makes sense. Many really love coding, and some lose sight that their role is to provide solutions and innovations.”
The reality is AI and ML can’t be applied to every scenario. For example, one company created a sophisticated ML approach to examining marketing analytics and performance so AI could predict the most effective marketing channels, but it required an incredible number of data integrations, and it wasn’t predictably better at suggesting marketing channels than existing experts using their expertise. Likewise it is not yet possible to train enough scenarios of kids running in the street or a car swerving into your lane for self driving cars to learn and act accordingly. What is the safe default? Should the car stop when it doesn’t know what to do? Should it revert to a manual mode? These are just some of the ongoing challenges in training the ‘last 10%” of AI despite the vast majority of driving decisions being more efficient and accurate than a human driving on the road today. Beyond the more obvious self-driving car scenarios, many AI and ML use cases bring ethical considerations that should be taken seriously. Like any other technology at scale, there must be frameworks and guardrails to help understand potential impact, mitigation paths, and when to forgo the use of these technologies altogether.
Even if the server is based in California and stores only Indian data, it does not come within the sovereign jurisdiction of India or Indian courts. While it may be technically possible to isolate a set of data within the server that is deemed perverse or critical evidence and electronically “seal” it, the California company might not be interested in blocking large space within its server or servers for which it has invested millions, for a court case in faraway India. This is just one example of the limits of national laws versus the limitless, borderless movements of data, made possible by technology. Hence, data localisation is an issue that has been a hot topic for governments around the world. Alongside comes the issue of data portability. What does the right to data portability mean? This is a right that allows anybody who has put a set of data into one service or site, to obtain it from that service or site and reuse it for their own purposes across different services. The sense of portability is in the moving, copying and transferring personal data from one to another service without compromising on security. This right will also incorporate the right to have the quality of data undiminished or unchanged. Within this right is also incorporated the caveat that all such data will have only been gathered from the user with his or her consent.
“If one country becomes more aggressive, then the equilibrium response is that all countries are going to end up becoming more aggressive,” says Alexander Wolitzky, an MIT economist who specializes in game theory. “If after every cyberattack my first instinct is to retaliate against Russia and China, this gives North Korea and Iran impunity to engage in cyberattacks.” But Wolitzky and his colleagues do think there is a viable new approach, involving a more judicious and well-informed use of selective retaliation. “Imperfect attribution makes deterrence multilateral,” Wolitzky says. “You have to think about everybody’s incentives together. Focusing your attention on the most likely culprits could be a big mistake.” The study is a joint project, in which Sandeep Baliga, the John L. and Helen Kellogg Professor of Managerial Economics and Decision Sciences at Northwestern University’s Kellogg School of Management added to the research team by contacting Wolitzky, whose own work applies game theory to a wide variety of situations, including war, international affairs, network behavior, labor relations, and even technology adoption.
As cars are becoming more autonomous, cities are becoming more intelligent by using more sensors and instruments. To drive this intelligence forward, smart city IT infrastructure must be able to capture, store, protect and analyze data from autonomous vehicles. Similarly, autonomous vehicles could greatly improve their performance by integrating data from smart cities. In smart city planning, stakeholders must consider how they will enable the sharing of data in both directions, to and from autonomous vehicles, and how that data can be analysed and acted upon in real-time, so traffic keeps moving and drivers, passengers and pedestrians are kept safe. This means that a city needs physical infrastructure to handle the growing numbers of autonomous vehicles that will be on the streets and an IT infrastructure that can easily manage data storage, performance, security, resilience, mobilisation and protection from a central management console. For example, there’s a case to be made that cities should already be building networks of smart sensors along the roadside. These would have the capability to measure traffic conditions and potentially even monitor obstacles such as fallen trees, traffic collisions or black ice.
Automation itself isn’t a problem, but without a reskilling strategy it will be. Here’s how quality non-degree credentials can help. This fear of automation is not new. As the late Harvard professor Calestous Juma laid out in his seminal book Innovation and Its Enemies: Why People Resist New Technologies, technological progress has always come with some level of public concern. The bellhops feared automatic elevators and so did bowling pin resetters. Video did indeed “kill the radio star” and it wasn’t long before internet media streaming services made video retailers obsolete in the mid-2000s. This “creative destruction” means that automation-enabling technologies will destroy jobs, but they will also increase productivity, lower prices and create new (hopefully better) jobs too. Some have even advocated that in order to help low-income workers, we should speed up the automation of low-income jobs. Non-degree credentials can help workers adapt. Automation can change the world for the better, but only we if prepare for it. To be sure, non-degree credentials are no silver bullet to automation displacement. A number of policy recommendations can help our world transition to new, high-quality jobs.
In the last decade, the Industrial Internet of Things (IIoT) has revolutionised predictive maintenance. Sensors record operational data in real-time and transmit it to a cloud database. This dataset then feeds a digital twin, a computer-generated model that mirrors the physical operation of each machine. The concept of the digital twin has enabled manufacturing companies not only to plan maintenance but to get early warnings of the likelihood of a breakdown, pinpoint the cause, and run scenario analyses in which operational parameters can be varied at will to understand their impact on equipment performance. Several eminent ‘brand’ products exist to create these digital twins, but the software is often challenging to customise, cannot always accommodate the specific needs of every manufacturing environment, and significantly increases the total cost of ownership. ML-powered digital twins can address these issues when they are purpose-built to suit each company’s specific situation. They are affordable, scalable, self-sustaining, and, with the right user interface, are extremely useful in telling machine operators the exact condition of the equipment under their care.
What does this mean for the future of AI research? Things may not be as bleak as they look. The cost of training might come down as more efficient training methods are invented. Similarly, while data center energy use was predicted to explode in recent years, this has not happened due to improvements in data center efficiency, more efficient hardware and cooling. There is also a trade-off between the cost of training the models and the cost of using them, so spending more energy at training time to come up with a smaller model might actually make using them cheaper. Because a model will be used many times in its lifetime, that can add up to large energy savings. In my lab’s research, we have been looking at ways to make AI models smaller by sharing weights or using the same weights in multiple parts of the network. We call these shapeshifter networks because a small set of weights can be reconfigured into a larger network of any shape or structure. Other researchers have shown that weight-sharing has better performance in the same amount of training time. Looking forward, the AI community should invest more in developing energy-efficient training schemes.
Quote for the day:
"Any man who has ever led an army, an expedition, or a group of Boy Scouts has sadism in his bones." -- Tahir Shah