MIT News recently highlighted the work of Audrey Lorvo, a senior at MIT who has been researching artificial intelligence (AI) safety. Her focus is on ensuring AI systems remain reliable, aligned with human values, and beneficial to society. Lorvo’s research explores both technical challenges—such as robustness and alignment—and broader concerns like transparency and accountability. Through her studies and participation in AI governance initiatives, she is helping to develop strategies that mitigate risks while supporting AI’s continued advancement.
Image Credit: Anggalih Prasetya/Shutterstock.com
The Growing Need for AI Safety Research
As AI capabilities continue to evolve at a rapid pace, concerns about safety and alignment with human values have become more pressing. AI safety research aims to ensure that these systems function as intended and do not pose risks to humanity. A key aspect of this work is AI alignment, which focuses on making AI decision-making processes reflect human goals. While previous research has explored robustness, interpretability, and value alignment, challenges remain in ensuring AI systems act in accordance with human intentions.
With the increasing possibility of artificial general intelligence (AGI), researchers stress the need for governance strategies and technical safeguards. Audrey Lorvo, a senior at MIT, has contributed to this field by examining AI automation in research and its broader societal implications. Her work focuses on closing the gaps in AI governance and technical safety, ensuring AI development remains both beneficial and controllable.
AI Alignment and Governance
One of the fundamental challenges in AI safety is ensuring AI systems align with human values and operate as intended. As AI models grow increasingly complex and autonomous, this task becomes even more demanding. Lorvo’s research explores how AI could accelerate its own development and the governance implications of such advancements. By analyzing these trends, she aims to identify strategies that can prevent unintended consequences before they arise.
As a scholar at MIT’s Schwarzman College of Computing Social and Ethical Responsibilities of Computing (SERC), Lorvo has examined the risks associated with AI automation in research and development. A key concern is AI optimizing processes in ways that are not fully understood by humans, potentially leading to unpredictable outcomes. Collaborating with experts, she has been developing governance strategies to ensure AI remains beneficial while maintaining necessary oversight.
Lorvo further expanded her expertise through the AI Safety Technical Fellowship, where she conducted in-depth analyses of AI alignment and governance research. This experience provided her with valuable insights into how organizations can implement effective policies. She has been a strong advocate for balancing innovation with safety, emphasizing the importance of transparency and accountability in AI development.
The Broader Impact: AI’s Social and Economic Influence
Beyond the technical challenges, AI is reshaping industries, job markets, and global economies. While it does have huge potential, rapid advancements raise ethical concerns, including fairness, accountability, and societal disruptions. Lorvo’s work bridges the gap between technical AI safety measures and their real-world applications.
With a background in computer science, economics, and data science, she examines AI from multiple perspectives. Her research explores how AI-driven automation may impact labor markets, contribute to economic disparities, and challenge existing governance frameworks. By collaborating with policymakers, legislators, and strategic advisors, she contributes to shaping AI policies that support responsible and equitable development.
Her involvement in MIT’s AI Alignment group has further strengthened her understanding of the intersection between AI safety and public policy. She advocates for data-driven AI governance frameworks that can adapt to technological advancements. Lorvo also underscores the importance of considering marginal impact—the additional effect of each decision—when formulating AI policies. This approach ensures that resources are allocated efficiently to maximize benefits while mitigating risks.
Looking Ahead
As AI continues to advance, making sure it remains safe and aligned with human values is more important than ever. Lorvo’s research is helping to bridge the gap between technical advancements and sound governance strategies, ensuring AI remains a tool that benefits society rather than one that operates beyond our control.
She is actively contributing to discussions on responsible AI development, focusing on alignment, transparency, and its economic impact. Her work highlights the need for a balance between technical expertise and thoughtful policy, shaping a future where AI is not just innovative but also ethical and responsible.
Disclaimer: The views expressed here are those of the author expressed in their private capacity and do not necessarily represent the views of AZoM.com Limited T/A AZoNetwork the owner and operator of this website. This disclaimer forms part of the Terms and conditions of use of this website.
Source:
Massachusetts Institute of Technology.