Page 115 - Cyber Defense eMagazine January 2023
P. 115

AI, by its nature, lends itself to breaking down data silos, as it streamlines data governance and assists
            with expediting the approval process for agencies sharing data. Still, the lack of fundamental legislation
            leads to separate AI requirements for each agency and makes cross-agency collaboration more difficult.

            For example, the Department of Energy may require a level of AI transparency that the Department of
            Homeland Security's AI cannot provide, making collaborating on a project difficult and creating silos of
            information that cannot be shared between the two agencies. Additionally, the network in which one AI
            system operates may look completely different from the other, making it even more challenging to share
            information.

            Actionable  legislation  will  help  solve  this  issue.  By  enforcing  government-wide  AI  regulations  and
            recommendations, agencies can work within each other's standardized AI networks with confidence that
            their guidelines are being met, breaking down the data silos created by different frameworks.

            However, standardized legislation may be more easily said than done.



            A World of AI Regulation

            No single framework or legislation can fulfill the mission requirements of all agencies, as AI is often
            mission-oriented.

            Take one critical aspect of AI, ethics, as an example. The AI Bill of Rights does feature guidelines on
            mitigating  discrimination  in  AI.  However,  agencies  have  different  ethical  considerations  and  risks  to
            consider. For instance, while the Department of Defense (DoD) must deal with life-and death-decisions
            for warfighters overseas, the Department of Education must look at student application bias or curriculum
            prejudice.

            However,  this  doesn't  mean  AI  legislation  can  ignore  the  issue  of  ethics.  Instead,  comprehensive
            legislation should showcase and clarify the plurality of AI by requiring each agency create a framework
            explicitly designed for their goals while still meeting baseline government-wide criteria.

            Incorporating language requiring developers consider the general challenges most AI solutions must
            address – such as bias, user safety, and implementation – while allowing a level of flexibility within the
            details can account for specific agency missions.

            For example, while both DoD and the Department of Education have different ethical considerations,
            guidelines such as "AI must only be used to support the safety and development of U.S. residents and
            citizens" apply to both agencies. Furthermore, requiring detailed guidelines be approved by the legislative
            branch ensures that each framework is viable and in line with these basic considerations.

            NIST has already laid the groundwork for this process by outlining several considerations in its AI Risk
            Management Framework, listing the characteristics of trustworthy systems as "valid and reliable, safe,
            fair  and  bias  is  managed,  secure  and  resilient,  accountable  and  transparent,  explainable  and
            interpretable, and privacy-enhanced." By adding a legal aspect to this list and elaborating on additional
            considerations, the government can create a foundation for AI legislation.






            Cyber Defense eMagazine – January 2023 Edition                                                                                                                                                                                                       115
            Copyright © 2023, Cyber Defense Magazine. All rights reserved worldwide.
   110   111   112   113   114   115   116   117   118   119   120