Making AI and information more Helpful, Humane, Harmless

Making Policy makers know the importance of ai safety

Learn More

The public policy on how these AI systems are developed will dictate how our future generations will use this technology.

Our Mission

  • Better interpretability
  • More robust systems
  • Scalable oversight
  • Good monitoring
  • Aligned Systems
Objectives

Our objective is to advance the AI race, document organizational risks, find and mitigate rogue actors, and make explainability of AI models more accessible to the public sector. We believe that having a clear understanding of internal workings of the neural nets is important and essential for the safe development of AI systems

  • Advance the frontier of red teaming efforts
  • Documentation and evangelizing interpretability efforts
  • creating rigorous conceptualizations and benchmarking efforts
  • Developing and sharing tools for AI safety
  • creating mechanisms to better understand the capabilities of different architectures
  • Developing standards for testing and benchmarking multi agent workflows
  • Developing better alignment and performance optimization

Consequences of Persuasive and Rogue agents

  • Ai based misinformation generation
  • Exploiting users trust and market manipulation
  • Centralized control of information
  • creating tailored content for targeted persuasion
“I’m increasingly inclined to think that there should be some regulatory oversight, maybe at the national and international level, just to make sure that we don’t do something very foolish. I mean with artificial intelligence we’re summoning the demon.” —Elon Musk, MIT’s AeroAstro Centennial Symposium
“I’m more frightened than interested by artificial intelligence – in fact, perhaps fright and interest are not far away from one another. Things can become real in your mind, you can be tricked, and you believe things you wouldn’t ordinarily. A world run by automatons doesn’t seem completely unrealistic anymore. It’s a bit chilling.”
—Gemma Whelan, the guardian

Concentration of power and cyber warfare

  • AI creating a totalitarian regime
  • Autonomous weaponry without human oversight
  • Destroying critical infrastructure like electrical grids, airline systems, cyber infra
  • AI driven extreme and irreversible concentration of power

Our Services

We specialize in providing comprehensive AI solutions. From AI safety testing to AI model evaluation, we offer a wide range of services to meet your needs. Our high-quality data repositories for evaluation ensure that your AI models are robust and reliable. Additionally, we advance the research frontier with our open-source AI research work.

Harms from Increasingly Agentic Systems

As AI systems become more agentic and increasingly interdependent, it becomes more difficult to align them with human values and intentions. This can lead to unintended consequences, such as harmful behaviors or misaligned goals.

Agentic Alignment
Security and Cyber Threats

AI systems have tremendous potential to be exploited by malicious actors. From adversarial attacks to model inversion, the risks are significant and complex. Consequences include autonomous vehicle failures, fraud, privacy violations, and more.

Security and Management
Concentration of Power

As AI systems advance, power dynamics will shift, creating monopolistic tendencies and increased persuasion for specific objectives.

Power Dynamics
Consequences of AI Race

The AI race drives innovation but risks deploying flawed systems, amplifying biases, privacy violations, and weapon misuse.

AI Race

Reliability


Most large-scale LLMs are trained on internet data, which is often factual but occasionally includes dishonest or biased responses. This can perpetuate misinformation and vulnerabilities in automated systems. We aim to develop large-scale datasets of misinformation to help LLM providers analyze, fine-tune, and mitigate risks, making systems more honest, helpful, and harmless.

  • Customized Solutions: We design and implement tailored solutions to meet your specific needs and find the right data for your model.
  • Data Democratization: Our mission is to make data accessible by creating repositories and crowdsourcing counterfactual responses.

AI Security & Governance


Our mission is to ensure policymakers understand AI safety’s importance and implement policies to develop these systems securely.

  • AI Policy: We prioritize data integrity and confidentiality to protect against breaches and misuse.
  • Our governance framework ensures AI systems are transparent, fair, and aligned with ethical standards.
AI Policy Illustration

Subscribe And Follow

Stay updated with our latest news and announcements. Subscribe to our newsletter and follow us on social media for valuable insights and exciting updates.