top of page
Buscar

Impactful research. At scale.

  • Foto del escritor: Matias Zabaljauregui
    Matias Zabaljauregui
  • 26 dic 2024
  • 1 Min. de lectura

Apart Research is a non-profit AI safety lab. We host open-to-all research sprints, publish papers, and incubate talented researchers to make AI safe and beneficial for humanity.



Safe AI

Publishing rigorous empirical work for safe AI: evaluations, interpretability and more.

“Do you have a design in mind for your blog? Whether you prefer a trendy postcard look or you’re going for a more editorial style blog - there’s a stunning layout for everyone.”

Novel approaches

Our research is underpinned by novel approaches focused on neglected topics.


Improving Llama-3-8B-Instruct Hallucination Robustness in Medical Q&A Using Feature Steering


This paper addresses the risks of hallucinations in LLMs within critical domains like medicine. It proposes methods to (a) reduce hallucination probability in responses, (b) inform users of hallucination risks and model accuracy for specific queries, and (c) display hallucination risk through a user interface. Steered model variants demonstrate reduced hallucinations and improved accuracy on medical queries. The work bridges interpretability research with practical AI safety, offering a scalable solution for the healthcare industry. Future efforts will focus on identifying and removing distractor features in classifier activations to enhance performance.


 
 
 

Comentarios


Venten.ai bridges the critical gap between enterprise AI utility and adversarial robustness. As organizations shift from static chatbots to autonomous agentic workflows, the attack surface expands from simple prompt injections to the hijacking of tools, memory, and decision-making logic. We provide the technical rigor necessary to ensure that your AI agents remain secure, aligned, and resilient against sophisticated cyber kill chains that target the very muscles and memory of your autonomous systems.

Our missionOur differential lies in a transition from traditional black-box testing to a research-driven, Defense-in-Depth methodology. By quantifying the Security Decay Factor in open-weight and quantized models, we implement specialized architectural layers—including semantic firewalls, dynamic output sanitization, and autonomous red teaming. At Venten.ai, we do not just test prompts; we audit the entire agentic lifecycle, transforming AI safety from a static benchmark into a dynamic, measurable KPI for high-stakes production environments.

bottom of page