
Human-Centric Concept AI: Deepika Vemuri on Interpretable Models
We often celebrate headline accuracy numbers for AI models – and then hand those opaque models to clinicians, judges, or farmers and ask them to explain a decision. That gap between performance and trust is where projects that prioritise interpretability become strategically important, not just academically interesting.
I recently came across an interview with Deepika Vemuri (IIT Hyderabad), whose PhD work explores two complementary directions in Concept-Based Learning: (1) replacing flat, linear concept-to-class mappings with learned, differentiable logic predicates; and (2) aligning concept learning with the depth-wise hierarchy of neural networks using Formal Concept Analysis. These may sound academic, but they point to three operationally relevant shifts for product teams and CTOs building real-world AI.
Why this matters for enterprise architecture
– From explanation to actionable intervention: Logic-based concept models (LogicCBMs) move beyond “these features contributed X%” to human-readable predicates (e.g., A AND (B XOR C)). That lets domain experts test counterfactuals and toggle misleading concepts – a form of surgical intervention that is far easier to operationalise than ad-hoc feature-attribution. For regulated or high-stakes domains, this capability transforms an explanation into an auditable remediation step.
– Robustness over average accuracy: The introduction of a worst-case metric – measuring how model confidence changes when misleading concepts are removed – highlights a critical product metric often missed in model evaluations. Enterprises should measure not only mean accuracy but robustness to spurious correlations and failure modes that matter to users.
– Modular, hierarchical concepts improve reuse and transfer: Mapping general concepts to earlier layers and specific concepts to deeper layers mirrors how complex systems are built: modules that are reusable and easier to test. This aligns with software architecture principles (separation of concerns, loose coupling) and makes transfer learning across tasks cheaper and safer.
Trade-offs and practical decisions
– Interpretability vs raw throughput: Logic predicates and structured concept hierarchies may add modelling complexity and annotation overhead. Expect longer model development cycles, but payback comes in reduced downstream risk, easier debugging, and faster regulatory approvals.
– Build vs buy: For commodity recommendation or perception tasks where transparency isn’t required, off-the-shelf black-box models may suffice. For healthcare, legal, public-sector, or safety-critical automation, adopt a hybrid approach: use a performant backbone for perception, but front it with a concept-based module to provide explanations and actionable interventions.
– Data and annotation strategy: Concept-based approaches require curated concept annotations and thoughtful hierarchy design. Invest early in taxonomy design and tooling for human-in-the-loop labelling – this is infrastructure work that pays dividends in model maintainability.
How to start (practical steps for CTOs and Founders)
– Add concept-level tests to your MLOps pipeline: unit-test concepts, assert expected predicate behaviour, and measure the worst-case metric alongside accuracy.
– Prototype concept overlays on top of existing models: create an interpretability layer that maps model features to human concepts before committing to full re-architecting.
– Engage domain experts for predicate validation, not just label collection: their input is essential for building meaningful logic combinations and for operational interventions.
– Plan for auditability: log concept activations and intervention experiments. These become evidence in audits and governance reviews.
A note for India and the Northeast context
This approach has direct relevance where trust and explainability matter – for example, crop advisory systems, telemedicine, or welfare-disbursement classifiers used by government agencies. Hierarchical concepts and logical predicates can reduce the risk of biased misclassification driven by spurious visual cues and make models usable by local experts who need to understand “why” a decision was made.
Closing thought
As AI moves from lab benchmarks into the hands of citizens and officials, the acid test is not whether a model is clever, but whether it can be inspected, reasoned about, and corrected by people who are accountable for its outcomes. Concept-based, logic-aware architectures offer a practical path toward AI systems that are both capable and trustworthy.
About the Author
Sanjeev Sarma is the Founder Director of Webx Technologies Private Limited, a leading Technology Consulting firm with over two decades of experience. A seasoned technology strategist and Chief Software Architect, he specializes in Enterprise Software Architecture, Cloud-Native Applications, AI-Driven Platforms, and Mobile-First Solutions. Recognized as a “Technology Hero” by Microsoft for his pioneering work in e-Governance, Sanjeev actively advises state and central technology committees, including the Advisory Board for Software Technology Parks of India (STPI) across multiple Northeast Indian states. He is also the Managing Editor for Mahabahu.com, an international journal. Passionate about fostering innovation, he actively mentors aspiring entrepreneurs and leads transformative digital solutions for enterprises and government sectors from his base in Northeast India.

