scorecardresearch
Tuesday, July 22, 2025
YourTurnSubscriberWrites: Responsible Scaling Policies (RSPs) for Generative AI–Beyond the AGI threat

SubscriberWrites: Responsible Scaling Policies (RSPs) for Generative AI–Beyond the AGI threat

LLMs show great promise for AGI but also pose risks. Safety frameworks like ASLs, continuous monitoring, and addressing disinformation are crucial for safer AI development.

Thank you dear subscribers, we are overwhelmed with your response.

Your Turn is a unique section from ThePrint featuring points of view from its subscribers. If you are a subscriber, have a point of view, please send it to us. If not, do subscribe here: https://theprint.in/subscribe/

Large Language Models or LLMs have been at the forefront of AI innovation for the last 3 years now and the most likely technological contenders for a leading us into a world with Artificial General Intelligence (AGI). 

Given the immense promise offered by any new technology it does not come without its own inherent risks. I can speak from personal experience that the models having gotten significantly better since I have started tracking them as a data science and AI professional

Although we hear that the models are beginning to saturate with the previous assumption of training the models with larger datasets – no longer holding as much value proposition as it did until recently

We know from the previous research and developments like – “instruction-based fine-tuning and more recently Agentic AI systems” will make the models more capable and lead to further advancements

Therefore, it is imperative that we think about improving the safety around these models and in this regard, I borrow heavily from Anthropic’s self-regulation framework on Responsible Scaling Policies (RSPs)

How can we understand and quantify safety for Large Language Models?

Anthropic measures safety levels in tiered manner called AI Safety Levels (ASLs)

  • ASL1 — This level is applicable to smaller models which are usually not assumed to pose a catastrophic risk
  • ASL2 — This is the level most publicly available models currently are assumed to pose. Think — Claude Sonnet 3.5, GPT-4/4o
  • ASL3 — At this level the risks are assumed to be much higher specifically posing risks such as cyberattacks, deliberate misuse by non-state actors
  • ASL4 — At the moment this is a speculative level beyond ASL-3 where can assume misuse by state actors or autonomy risks for the model

Conclusion – What else is needed to make a world with safer AI systems?

  • AI safety is currently measured against human benchmarks that are continuously being surpassed by models that are achieving human-like competencies on specific tasks
  • Ensuring the compliance of AI Safety levels across global organizations — state and non-state
  • Continuous monitoring of developed model’s pre-release in an open and transparent manner to ensure rigorous testing before launch
  • Addressing concerns on the use of these systems for the spread of disinformation/mis-information

These pieces are being published as they have been received – they have not been edited/fact-checked by ThePrint. 

Subscribe to our channels on YouTube, Telegram & WhatsApp

Support Our Journalism

India needs fair, non-hyphenated and questioning journalism, packed with on-ground reporting. ThePrint – with exceptional reporters, columnists and editors – is doing just that.

Sustaining this needs support from wonderful readers like you.

Whether you live in India or overseas, you can take a paid subscription by clicking here.

Support Our Journalism

LEAVE A REPLY

Please enter your comment!
Please enter your name here