AI Society Tracker

Abigail Yohannes

My project aimed to develop a platform for real time and democratized data on ai in society

Reviewer's Comments

Reviewer's Comments

Arrow
Arrow
Arrow

Ziba Atak

Strengths:

- Novel Idea: The concept of a real-time AI societal impact tracker is innovative and addresses a critical gap in AI safety and policy.

-Relevance: The paper highlights the importance of tracking AI’s societal impact, particularly in areas like job markets, ethics, and policymaking.

-Groundwork: A strong foundation for future work, including identifying data sources and outlining the need for such a tool.

Areas for Improvement:

-Methodology and Documentation: The paper lacks a methodology section and detailed documentation (due to time constraints), making it difficult to evaluate the technical implementation. Include a clear explanation of how the tracker works and provide access to the codebase.

-Results and Insights: The results section is missing due to time constraints, and the app does not provide enough information to interpret the data. Include a detailed analysis of the findings and their implications.

-Limitations and Risks: The paper does not thoroughly analyze the limitations of the tracker itself or the potential risks of implementing such a tool. Expand on these aspects to strengthen the analysis.

-Scalability and Generalization: Demonstrate how the tracker could be scaled or generalized to broader applications. Include a discussion of future steps and scalability.

Suggestions for Future Work:

-Develop a detailed methodology and document the technical implementation, including access to the codebase.

-Conduct a thorough analysis of the results and their implications for AI safety and policy.

-Explore mitigation strategies for the risks and limitations.

-Investigate the tracker’s performance in diverse contexts and expand the data collection process.

Nakshathra Suresh

The submission draws upon quite a few novel ideas for their 'AI Societal Impact Tracker' and should be appreciated for their innovative idea. The paper was also well written, and uses formal language in a clear, presentable manner. However, in order to strengthen the submission, greater usage of literature should have been applied. Unfortunately the lack of depth in the sources means that there is not enough literature as a foundation. There is plenty of existing research that surrounds the need to examine the impact of AI on society, therefore it would be recommended that in the future, the submission draws upon more research and statistics. The paper is also incomplete, which means there is less discussion and analysis that can be assessed for the submission. Overall a great effort.

Andreea Damien

Hi Abigail, I wanted to say that I appreciate the fact that you decided to submit your project despite it not being finished. Overall, I think the idea is great and very practically relevant, and something that you should definitely pursue if you’re passionate about. As to the project itself, you could use what you’ve written so far as a ‘call for action’ type of blogpost, or develop it further and publish as a peer-reviewed ‘perspective’ article. We can also help you further, if you want to build the actual platform.

Cite this work

@misc {

title={

@misc {

},

author={

Abigail Yohannes

},

date={

3/10/25

},

organization={Apart Research},

note={Research submission to the research sprint hosted by Apart.},

howpublished={https://apartresearch.com}

}

May 20, 2025

EscalAtion: Assessing Multi-Agent Risks in Military Contexts

Our project investigates the potential risks and implications of integrating multiple autonomous AI agents within national defense strategies, exploring whether these agents tend to escalate or deescalate conflict situations. Through a simulation that models real-world international relations scenarios, our preliminary results indicate that AI models exhibit a tendency to escalate conflicts, posing a significant threat to maintaining peace and preventing uncontrollable military confrontations. The experiment and subsequent evaluations are designed to reflect established international relations theories and frameworks, aiming to understand the implications of autonomous decision-making in military contexts comprehensively and unbiasedly.

Read More

Apr 28, 2025

The Early Economic Impacts of Transformative AI: A Focus on Temporal Coherence

We investigate the economic potential of Transformative AI, focusing on "temporal coherence"—the ability to maintain goal-directed behavior over time—as a critical, yet underexplored, factor in task automation. We argue that temporal coherence represents a significant bottleneck distinct from computational complexity. Using a Large Language Model to estimate the 'effective time' (a proxy for temporal coherence) needed for humans to complete remote O*NET tasks, the study reveals a non-linear link between AI coherence and automation potential. A key finding is that an 8-hour coherence capability could potentially automate around 80-84\% of the analyzed remote tasks.

Read More

Mar 31, 2025

Model Models: Simulating a Trusted Monitor

We offer initial investigations into whether the untrusted model can 'simulate' the trusted monitor: is U able to successfully guess what suspicion score T will assign in the APPS setting? We also offer a clean, modular codebase which we hope can be used to streamline future research into this question.

Read More

May 20, 2025

EscalAtion: Assessing Multi-Agent Risks in Military Contexts

Our project investigates the potential risks and implications of integrating multiple autonomous AI agents within national defense strategies, exploring whether these agents tend to escalate or deescalate conflict situations. Through a simulation that models real-world international relations scenarios, our preliminary results indicate that AI models exhibit a tendency to escalate conflicts, posing a significant threat to maintaining peace and preventing uncontrollable military confrontations. The experiment and subsequent evaluations are designed to reflect established international relations theories and frameworks, aiming to understand the implications of autonomous decision-making in military contexts comprehensively and unbiasedly.

Read More

Apr 28, 2025

The Early Economic Impacts of Transformative AI: A Focus on Temporal Coherence

We investigate the economic potential of Transformative AI, focusing on "temporal coherence"—the ability to maintain goal-directed behavior over time—as a critical, yet underexplored, factor in task automation. We argue that temporal coherence represents a significant bottleneck distinct from computational complexity. Using a Large Language Model to estimate the 'effective time' (a proxy for temporal coherence) needed for humans to complete remote O*NET tasks, the study reveals a non-linear link between AI coherence and automation potential. A key finding is that an 8-hour coherence capability could potentially automate around 80-84\% of the analyzed remote tasks.

Read More

May 20, 2025

EscalAtion: Assessing Multi-Agent Risks in Military Contexts

Our project investigates the potential risks and implications of integrating multiple autonomous AI agents within national defense strategies, exploring whether these agents tend to escalate or deescalate conflict situations. Through a simulation that models real-world international relations scenarios, our preliminary results indicate that AI models exhibit a tendency to escalate conflicts, posing a significant threat to maintaining peace and preventing uncontrollable military confrontations. The experiment and subsequent evaluations are designed to reflect established international relations theories and frameworks, aiming to understand the implications of autonomous decision-making in military contexts comprehensively and unbiasedly.

Read More

Apr 28, 2025

The Early Economic Impacts of Transformative AI: A Focus on Temporal Coherence

We investigate the economic potential of Transformative AI, focusing on "temporal coherence"—the ability to maintain goal-directed behavior over time—as a critical, yet underexplored, factor in task automation. We argue that temporal coherence represents a significant bottleneck distinct from computational complexity. Using a Large Language Model to estimate the 'effective time' (a proxy for temporal coherence) needed for humans to complete remote O*NET tasks, the study reveals a non-linear link between AI coherence and automation potential. A key finding is that an 8-hour coherence capability could potentially automate around 80-84\% of the analyzed remote tasks.

Read More

May 20, 2025

EscalAtion: Assessing Multi-Agent Risks in Military Contexts

Our project investigates the potential risks and implications of integrating multiple autonomous AI agents within national defense strategies, exploring whether these agents tend to escalate or deescalate conflict situations. Through a simulation that models real-world international relations scenarios, our preliminary results indicate that AI models exhibit a tendency to escalate conflicts, posing a significant threat to maintaining peace and preventing uncontrollable military confrontations. The experiment and subsequent evaluations are designed to reflect established international relations theories and frameworks, aiming to understand the implications of autonomous decision-making in military contexts comprehensively and unbiasedly.

Read More

Apr 28, 2025

The Early Economic Impacts of Transformative AI: A Focus on Temporal Coherence

We investigate the economic potential of Transformative AI, focusing on "temporal coherence"—the ability to maintain goal-directed behavior over time—as a critical, yet underexplored, factor in task automation. We argue that temporal coherence represents a significant bottleneck distinct from computational complexity. Using a Large Language Model to estimate the 'effective time' (a proxy for temporal coherence) needed for humans to complete remote O*NET tasks, the study reveals a non-linear link between AI coherence and automation potential. A key finding is that an 8-hour coherence capability could potentially automate around 80-84\% of the analyzed remote tasks.

Read More

This work was done during one weekend by research workshop participants and does not represent the work of Apart Research.
This work was done during one weekend by research workshop participants and does not represent the work of Apart Research.