Securing Critical Infrastructure
govciooutlookeurope

Securing Critical Infrastructure

Government CIO Outlook | Tuesday, January 17, 2023

Securing critical infrastructure is essential for our global economy and society.

FREMONT, CA: Events that may weaken the confidentiality, integrity or availability of the services delivered by crucial infrastructure providers and their networks could have important and potentially devastating consequences. Certainly, governments are increasingly focused on this problem. As a result, they are calling for critical infrastructure providers and their IT vendors to implement technical and organizational security measures and prepare for the potential impacts of security incidents.

Stay ahead of the industry with exclusive feature stories on the top companies, expert insights and the latest news delivered straight to your inbox. Subscribe today.

Qualifying trustworthy IT vendors

Evaluating the entire practices of a vendor's organization should be the initial point. That incorporates assessing the robustness, repeatability and consistency of their secure development practices and transparency about vulnerabilities detected in their products, which is vital for resilience.

While evaluating a point solution is a step in the correct direction, a holistic approach that considers the function of people, processes and technology in protecting critical global infrastructure will yield a far better result. Furthermore, point-product security is fleeting and unreliable if the organization producing the solution lacks the process maturity to consistently demonstrate its trustworthiness.

Security does not end when a vendor places a solution on the market. How a critical infrastructure worker architects, deploys, monitors and keeps its networks and information systems on an ongoing basis is important to secure operations. An active security architecture that is resilient and trustworthy will help prevent, detect and react to cyber threats.

Reliable solutions are products or services that do what is awaited in a verifiable way. Vendors can build security capabilities into technologies at the design phase. These incorporate validation of crypto modules; image signing to create special digital signatures that can be checked at runtime; hardware-anchored secure boot to spontaneously verify software integrity at boot-up; technologies and processes to confirm that the hardware is genuine; and runtime defenses that help protect against injection attacks of malicious code into running software. Moreover, vendors must know what is in their code and why it's there; doing so is fundamental to a mature and secure engineering process.

Vendors can also support network operators in verifying the integrity of their technology once it's deployed in a network operation. But, again, corroborating that the infrastructure hardware and software are working as expected is the key to maintaining the architectural components' good security posture and integrity.

Qualifying secure solutions

Revising procurement regulations to command better assessment of vendor solutions is now delayed. Government regulations should need that any technology deployed in critical infrastructure be procured only from provably trustworthy vendors.

Derive that proof from mandatory security assessments. Instead, start by leveraging baseline measures of adherence to simple security measures already captured in internationally recognized standards like Common Criteria. These are beneficial as a starting point and can serve as appropriate yardsticks for technology deployed broadly in less critical networks.

For mission-critical networks, extensive security assessments should be carried out by recognized, trusted experts. This may involve government agencies performing the testing themselves to ensure the results' quality and the shortage of skilled experts. Testing might also be performed with the support of select, highly qualified testing labs.

This can't be accosted as a mere compliance exercise, as it has become commonplace when assessing basic security standards. Robust security assessments directed at critical networks should employ vigorous and dynamic vetting of numerous critical vendor capabilities:

• Source code verification

• Design documentation

• Actual penetration-style solution testing

• The testing of artifacts and other relevant materials

Escort the assessment to an agreed-upon, secure location where the vendor's intellectual property will be protected.

Be certain the testing procedure keeps pace with market innovations and integrates a rigorous, risk-based approach. To allow efficiency, scale and expediency:

i) Manage product iterations by restricting testing to the updated part of a build. This will overcome the cost and time-to-market implications of testing every version.

ii) Build on proven assessment examples instead of beginning from scratch. Upgrade only when meaningful and collective value can be included.

iii) Collaborate with like-minded governments to build toward mutual recognition of testing, centering on mitigating cyber-risk rather than adhering to local business customs. This will decrease fragmentation across borders and enhance each country's ability to effectively scale its efforts.

Qualifying responsible operations

Emigrating to digital capabilities requires critical infrastructure providers to keep up with the latest threat monitoring and detection technologies. For example, machine-learning algorithms can help detect anomalies from the normal network and user behavior. That data can then be employed for informing control-based policies to mitigate attacks.

The vendor assists the infrastructure provider in deploying and operating their technology most effectively and securely. As operators require tools for onboarding and managing devices, vendors should work with them to guarantee that devices can be tested, provisioned and revised securely. Granting unique device identities, validated at set-up, is just one step in how this could be approached.

Asset, patch and vulnerability management are essential to the total lifecycle management of the security architecture and its elements. Therefore, IT vendors must track a strict process for managing security exposure information related to their solutions and networks.

Infrastructure providers will greatly advantage from requiring transparent and predictable approaches to vendors' vulnerability management and disclosures. That comprises published guidelines for timely vendor action to provide necessary patches.

It's important to patch and improve proactively and not wait until something bad happens.

Verify before trust

Words of confidence are not enough; vendors must demonstrate a range of behaviors that demonstrate they are a trusted partner and then incorporate those behaviors consistently throughout their operations.

With verification checkpoints in place, by working with rightly trusted vendors, and armed with the power of digital capabilities, our critical global infrastructure will be ready for the risks of tomorrow.

More in News

The landscape of modern policing is increasingly documented. From body-worn cameras (BWCs) and dashcams to stationary surveillance systems, law enforcement agencies are capturing an unprecedented volume of video data. This surge in digital evidence has brought a parallel rise in demand for public transparency, driven by open records laws and a community desire for accountability. However, this demand exists in a delicate balance with the fundamental right to privacy. Releasing raw footage is rarely an option, as it often contains sensitive and personally identifiable information (PII) of victims, witnesses, minors, and uninvolved bystanders. This is where video redaction—the process of obscuring sensitive information within video and audio files—becomes a critical, non-negotiable function. Over the past two decades, the methodology for performing this task has undergone a profound transformation, evolving from painstaking manual labor to a high-speed, technologically advanced process driven by artificial intelligence. The Manual Era: A Frame-by-Frame Bottleneck In the early days of digital video evidence, redaction was an entirely manual, labor-intensive endeavor. It was a task that fell to video technicians, investigators, or IT staff, requiring them to sit at a workstation with general-purpose video editing software. The process was granular and grueling. An analyst would load a video file and meticulously scrub through it, often frame by frame. Upon identifying a face, license plate, home address on a building, or a computer screen displaying personal data, they would have to apply an obscuring effect manually—typically a blur filter or an opaque black box. This wasn't a one-time "click and forget" action. The analyst had to "keyframe" the redaction, manually adjusting the box's position, size, and shape in subsequent frames to track the moving object or person. For a single 10-minute video clip featuring multiple individuals in a dynamic, unstable environment (like an officer walking through a crowd), this process could take many hours, sometimes even a full day's work. The workflow was linear, inefficient, and created a massive operational bottleneck. Agencies found themselves with a growing backlog of video evidence required for court discovery or public release, but with an equally growing deficit of person-hours to process it. The cost was not just in time and resources; it was in the significant potential for human error. Fatigue could easily cause an analyst to miss a face in a crowd or a reflection of a victim's ID in a window, leading to a critical privacy breach. The Catalyst for Change: A Deluge of Data The widespread adoption of body-worn cameras marked the tipping point. Suddenly, it wasn't just a few dashcam videos from specific incidents that needed processing. Agencies were now generating thousands of hours of footage every single day from hundreds of officers. The manual redaction model didn't just bend under this new weight; it broke. Simultaneously, the legal and social environment was shifting. Public records requests for BWC footage became routine, and court-mandated deadlines for evidence disclosure grew stricter. Agencies were caught between the public's right to know and the legal imperative to protect privacy. The sheer scale of the data made the old way impossible. It was clear that a technological leap was necessary to manage data flow, meet legal obligations, and maintain public trust. The Dawn of Automation: Machine Learning Takes the Wheel The solution emerged from the fields of artificial intelligence and machine learning. Instead of having a human manually find and track objects, new platforms were developed to automate this process. This shift from manual to machine learning represents the single most significant evolution in the history of video redaction. Modern redaction systems are powered by sophisticated computer vision models trained on vast datasets to identify specific objects with exceptional speed and accuracy. When an analyst uploads a video, the system automatically scans every frame, detecting and tagging relevant elements. Standard models are pre-trained to recognize common personally identifiable information (PII), such as faces, bodies, and license plates. At the same time, advanced systems can be customized to identify user-defined objects, such as agency-issued devices, tattoos, or credit cards. Once detected, the AI employs intelligent tracking to “lock on” to each object, applying the chosen redaction effect—blur, pixelation, or masking—throughout the video, even as the object moves or becomes partially obscured. The evolution of redaction technology extends to audio, with Natural Language Processing (NLP) enabling automatic transcription and searchable audio redaction. Analysts can quickly locate and censor sensitive terms such as names, addresses, or identification numbers without manually reviewing the entire recording. Significantly, automation enhances rather than replaces human oversight. The AI handles the labor-intensive tasks, producing a redacted draft within minutes, which human analysts then review for quality assurance—correcting minor errors or false positives as needed. This technological leap has transformed video and audio redaction from a time-consuming, error-prone process into a streamlined, efficient, and scalable workflow for law enforcement agencies. Agencies can now process and release video evidence in a fraction of the time, improving transparency and responsiveness to public records requests. Detectives and officers are freed from the tedious task of video editing, allowing them to focus on core investigative duties. Most importantly, the consistency and accuracy of machine learning reduce the risk of human error, providing greater protection for citizen privacy. The evolution from manual redaction to machine learning is more than just a technological upgrade. It is a foundational change that enables law enforcement to navigate the complex demands of the 21st century—balancing the critical needs for transparency, accountability, and the unassailable right to privacy in an increasingly documented world. ...Read more
Local government agencies are transitioning from outdated, fragmented legacy systems to adopt cloud-based software solutions. This shift goes beyond simply embracing new technology; it represents a significant modernization effort that improves municipal efficiency, lowers costs, and provides vastly enhanced services to citizens. Key Pillars of Cloud-Driven Efficiency The adoption of cloud-based software delivers measurable improvements across several dimensions of municipal operations. Cost optimization and scalability stand at the forefront of these benefits. By transitioning from large capital expenditures on hardware and perpetual licenses to a flexible, subscription-based “pay-as-you-go” model, municipalities can significantly reduce upfront costs and ongoing maintenance expenses. This shift eliminates the need for frequent hardware upgrades and inflated IT budgets. Moreover, cloud environments offer scalability, allowing municipalities to instantly adjust computing resources in response to demand fluctuations—whether during tax season, emergency responses, or permit application surges. This adaptability ensures efficient infrastructure utilization without the financial strain of maintaining excess capacity for peak loads. Equally transformative is the impact on service delivery and citizen experience. Cloud solutions enable 24/7 access to digital self-service portals where residents can submit service requests, apply for permits, pay bills, and track progress online—enhancing convenience, transparency, and satisfaction. Automating routine administrative workflows reduces manual data entry and approval bottlenecks, freeing municipal staff to focus on higher-value community services. For instance, e-permitting systems have been shown to reduce processing times from weeks to days, accelerating project delivery and boosting public confidence in local governance. Data, Security, and Applications in Modern Municipal Management Cloud technology also redefines how municipalities manage data and make decisions. By consolidating information across departments—such as planning, finance, and public safety—cloud platforms eliminate data silos and establish a single source of truth. This unified access fosters collaboration and consistency, while real-time analytics empower leaders to make data-driven decisions that optimize operations, from waste collection routing to traffic management and emergency planning. Cloud infrastructure also enhances security and resilience, offering levels of protection often beyond the reach of smaller municipalities. Leading cloud providers invest heavily in advanced encryption, continuous threat monitoring, and compliance with rigorous standards such as FedRAMP and NIST. These measures ensure robust data protection while mitigating cybersecurity risks. In the event of natural disasters or system outages, cloud-based disaster recovery mechanisms enable rapid restoration of services and data from any connected location—ensuring uninterrupted continuity of government functions. The practical impact of these technologies is evident across diverse municipal functions. In permitting and licensing, cloud-based portals and mobile inspection tools streamline applications and reduce human error. Enterprise asset management benefits from real-time GIS integration and lifecycle tracking, optimizing maintenance schedules and extending infrastructure lifespan. Financial management systems (ERP) unify budgeting, procurement, and payroll, offering real-time financial transparency and improved compliance. Similarly, utility billing and payments leverage automation and self-service tools to improve billing accuracy, minimize administrative workload, and enhance revenue collection efficiency. Together, these applications exemplify how cloud transformation is driving smarter, more responsive, and more efficient local governance. The migration to the cloud represents a strategic move for local governments to not only address the challenges of aging systems but also to lay the foundation for future innovation. This infrastructure is the bedrock for implementing AI for better forecasting, machine learning for fraud detection, and the Internet of Things (IoT) for smarter city management. While the transition requires careful planning, addressing data governance concerns, and ensuring staff training, the long-term gains in efficiency, cost savings, and the quality of citizen services make cloud-based software the indispensable engine for the modern municipal government. ...Read more
The increasing frequency of climate-related disasters, intensified cybersecurity threats, and rapid urbanization pose significant challenges for emergency management. Implementing innovative solutions that enhance resilience and improve response strategies is essential for effectively addressing these complex risks. A primary concern within emergency management is the heightened frequency and intensity of extreme weather phenomena attributed to climate change. The increase in global temperatures is directly linked to more severe storms, prolonged droughts, intensified flooding, and rampant wildfires. Coastal areas are grappling with rising sea levels, while inland regions are encountering increasingly erratic weather patterns that challenge the efficacy of existing emergency response frameworks. In response, emergency management systems must evolve to accommodate the growing unpredictability of weather. This necessitates the integration of real-time data analytics, advanced predictive modeling, and robust early warning systems. Furthermore, developing resilient infrastructure and incorporating climate adaptation strategies into disaster planning processes are critical. Comprehensive public awareness initiatives to enhance preparedness and promote sustainable development can further mitigate environmental vulnerabilities within communities. Governments and emergency management agencies must prioritize investments in climate resilience standards, ensuring that response systems are optimized to minimize the impact of future disasters. Collaboration among stakeholders, including public agencies, private sector entities, and community organizations, will be crucial in shaping effective and sustainable emergency management approaches in the face of escalating climate risks. Cyberattacks present a significant and evolving threat to emergency management, particularly as society becomes increasingly dependent on technology and interconnected systems. Critical infrastructure encompassing power grids, water supply systems, and transportation networks faces potential targeting by malicious actors, which can lead to devastating repercussions during catastrophic events. Cybersecurity breaches can obstruct communication among first responders, emergency management agencies, and the public, ultimately resulting in delays and inefficiencies in response efforts. Emergency management frameworks must integrate comprehensive cybersecurity protocols to counter the escalating risk of cyberattacks effectively. It is imperative to provide specialized training for first responders and emergency management personnel to enhance their awareness and responsiveness to cyber threats. Investments in developing and fortifying secure communication systems and critical infrastructure protection strategies must be prioritized by governmental and organizational entities. Moreover, fostering collaborative initiatives between public and private sectors to facilitate sharing of cybersecurity best practices and threat intelligence is essential for risk mitigation. Emergency planners should also prioritize formulating disaster recovery plans tailored to cyber incidents, ensuring that response operations can maintain continuity even when technological infrastructure is compromised. Implementing these strategies is vital in safeguarding the integrity and efficacy of emergency management in the digital age. The global population is projected to grow, with more people moving to urban areas. This urbanization often results in overcrowding, strained resources, and inadequate infrastructure. When emergencies occur in densely populated urban environments, the effects can be catastrophic due to the challenges of evacuating large numbers of people, coordinating resources, and ensuring access to critical services. Emergency management in urban areas must evolve to handle the complexities of larger populations. Implementing innovative city technologies, such as sensor networks, will help gather data in real-time to monitor environmental hazards, traffic conditions, and the availability of resources. Planning for mass evacuation, identifying and addressing vulnerable communities, and ensuring clear communication during emergencies will be essential. ...Read more
The relationship between citizens and the state is influenced not only by policies but also by a technological shift that has redefined expectations for service delivery across all sectors of society. In an age where the private sector provides instantaneous, personalized, and intuitive digital experiences, citizens now anticipate the same level of service from public institutions. The driving force behind this new era of governance is, without a doubt, cloud computing. More than just an IT infrastructure choice, cloud technology has become the essential platform on which responsive, resilient, and citizen-focused public services are built. The New Digital Social Contract Today’s citizens navigate their lives through smartphones, expect on-demand access to information, and value seamless, integrated experiences. This conditioning has forged a new, unspoken social contract: public services should be as accessible, reliable, and easy to use as the best consumer applications. The era of long queues, duplicative paperwork, and siloed departmental interactions is fading into obsolescence. The modern expectation is for a unified, proactive, and personalized relationship with government. Citizens envision a future where renewing a driver's license, registering a business, accessing healthcare records, or paying taxes can be accomplished through a single, secure digital portal, accessible at any time and from anywhere. They expect government agencies to know who they are, understand their needs based on previous interactions and life events, thereby personalizing the services offered. This demand for a consumer-grade experience is the primary catalyst compelling public sector bodies to reimagine their service delivery models from the ground up. The Architectural Foundation: Cloud-Native Elasticity and Agility At the most fundamental level, the cloud provides elasticity. Public services often experience fluctuating demand. Consider the surge in traffic on a tax portal during filing season, the massive data processing required for a national census, or the sudden need for a public health information hub during a crisis. In a traditional on-premise model, agencies would have to procure and maintain hardware for peak capacity, leaving vast resources underutilized most of the time. Cloud platforms eliminate this inefficiency. They offer a model of resource elasticity, where computational power, storage, and network bandwidth can be scaled up or down in near real-time. This can be represented by the principle of on-demand allocation, where ResourcesDeployed​∝DemandActual​. This ensures that services remain performant and available during peak loads while maintaining cost-efficiency during periods of regular activity. Beyond scalability, the cloud fosters unprecedented agility. Modern cloud-native development, utilizing principles such as microservices and Application Programming Interfaces (APIs), enables agencies to build, deploy, and update services with remarkable speed and agility. Instead of monolithic, slow-to-change systems, services are constructed as a collection of smaller, independent components. This modular approach enables the addition of new features to a mobile application or the reflection of policy changes in a benefits calculator in weeks or days, rather than months or years. APIs act as the connective tissue, enabling different systems and departments to securely share data and functionality, thereby breaking down the information silos that have historically hindered holistic service delivery. From Data Repositories to Intelligent Insights The cloud has fundamentally changed the government's relationship with data. Historically, data was often trapped within specific departments, stored in disparate formats, and challenging to aggregate for meaningful analysis. Cloud-based data platforms offer a unified environment for ingesting, storing, and processing vast quantities of information. This centralization creates the opportunity to move beyond simple record-keeping towards data-driven governance. By applying advanced analytics, machine learning, and artificial intelligence tools available on major cloud platforms, agencies can transform raw data into actionable intelligence. This capability allows for evidence-based policymaking, where real-time trends and predictive models inform decisions. Operationally, it enables the optimization of public resources, from managing traffic flow in smart cities to predicting maintenance needs for public infrastructure. For the citizen, it powers the delivery of proactive and predictive services. A system can, for example, automatically notify a family of their eligibility for a new childcare benefit upon registration of a birth, or alert a small business owner about a new grant they qualify for based on their industry and location. The ultimate trajectory of this evolution is the concept of "Government-as-a-Platform" (GaaP). In this model, the government provides the core, secure digital infrastructure—digital identity, secure payment gateways, and data-sharing APIs—upon which a rich ecosystem of public services can be built. This platform approach fosters innovation and enables the rapid development of new citizen-facing solutions. The citizen experience in a GaaP model is one of complete coherence. An individual interacts with a single digital identity that serves as their passport to all government services. This unified portal offers a personalized dashboard that displays relevant information and pending tasks, including upcoming vehicle inspections and voting registration deadlines. The experience is omnichannel, seamlessly moving between a web browser, a mobile app, and an intelligent chatbot, with the interaction context maintained across all channels. This forward-looking model is not a distant vision but the logical continuation of the current digital transformation. By leveraging the immense power of cloud computing, public institutions are progressively dismantling the barriers of the past. They are building services that are not only more efficient for the government, but more importantly, are more respectful of citizens’ time and needs. The journey is one of continuous iteration and improvement, moving public administration from a provider of static services to an orchestrator of intelligent and deeply human-centric outcomes. The cloud era is providing the tools not just to digitize government, but to reinvent it for a new generation. ...Read more

Weekly Brief