Features
University Education in the 21st Century
The basic principles that we should work on include –
a) Breadth of knowledge in context, not depth that is
essential only for those going on to do research
b) Better communication skills including teaching
skills that will facilitate the sharing of knowledge
c) Thinking skills that promote innovation and
analysis of different perspectives
d) Social awareness and sensitivity that contributes to
coherent and productive planning and action in the
world of work
———————-
Text of keynote address on ‘The Future of University Education’ delivered by Prof. Rajiva Wijesinha at the Sabaragamuwa Student Symposium, yesterday.
It is a pleasure to come back to Sabaragamuwa, and the more so this time as it is after several years. And though I am now old and lazy, and was a bit put out when I was told the text of my presentation was needed in advance, the topic given me, ‘University Education in the 21st Century’ was convenient. For I had been in fact reflecting on the subject, for the posts I now put up daily on what I term my political Facebook Account, that of the Council for Liberal Democracy.
The series that I called ‘Productive Initiatives’, as a contrast to the depressing series I am also writing, about the destruction wrought by J R Jayewardene and his political heirs, also however turned out depressing. For unfortunately most positive initiatives in education in the 20th century were promptly subverted, beginning with Kannangara’s Central Schools which J R Jayewardene soon straitjacketed in monolingualism.
It was the second great educational innovator of the last century, Prof Arjuna Aluwihare, who conceived the idea of Affiliated University Colleges, from which this University springs. He engaged in such innovation because by the eighties it had become clear to many youngsters that the education system was a mess. Though we prided ourselves on our literacy rates, and on providing free education up to university level, it was clear that the quality of the education provided was abysmal as far as many students were concerned and did not help them to gain decent employment.
There were of course many other causes for the radicalization of the young, and the insurgency that burst out at the end of the eighties. But the Youth Commission report that President Premadasa commissioned noted clearly the need to expand opportunities for rural youth. And though the school education system continued a mess, Arjuna Aluwihare as Chairman of the UGC proposed a radical new approach to tertiary education, and set up what were called Affiliated University Colleges, intended to provide a broad education to youngsters, including compulsory English and wider general awareness, instead of concentration on one or more subjects with no effort to relate them to the world of work.
His ideas were not supported by the majority of universities which were happy to continue doing the same thing for a few more decades. The only university which embraced the idea enthusiastically was Sri Jayawardenepura, which had a very dynamic Vice-Chancellor, Prof S B Hettiarachchi. So USJP conducted programmes in five AUCs, including the flagship programme of the AUCs, the English Diploma course.
This was open to students who had not done English at the Advanced Level. Very few schools in fact offered English at that level, so the intake at the three universities which offered English as a Special Subject was largely confined to students from Colombo and Kandy. This meant that only a few students offered English each year at these universities, but the Departments got vast amounts of funding on the grounds that they were producing English teachers for the nation. Given their exclusivity, hardly any of their products went into teaching except at a few elite schools.
I had long complained about restrictions on the study of English at tertiary level, and when I mentioned this to Arjuna he promptly got me involved in his new programme, which I had not known about before. And I believe I have every reason to be proud of what I achieved. One of my brightest colleagues at USJP, where I first returned to the University system, once told me that there was no point in being a teacher unless one’s pupils turned out better than one was oneself. When I look at my students who excel here, and at the Uva Wellassa University, and in the Department of Technical Education and Training, I feel that at least the English programmes I began in the early nineties have succeeded. Those I started later have done less well, but that is another story, once again to do with Ranil Wickremesinghe’s wickedness.
However, while the Science courses that Aluwihare initiated also I believe did well – and full marks again to this university in particular for that – the AUCs and their successors failed in other respects to live up to his vision. For instance the general courses he had thought of, essentially to increase the general knowledge of youngsters woefuly deprived of this in schools, fell into the trap that affected general courses elsewhere. They were stuffed full of specialist knowledge, and did not engage students to think of the realities of the world they lived in. Sri Lankan studies for instance regurgitated what students had learnt in school, without helping students to position Sri Lanka in the modern world. Unfortunately there was no clear understanding of soft skills, which we are now told at every turn is what the Sri Lankan education system fails to inculcate. And no one thought in those distant days of studying what happened elsewhere, of looking for instance at the development of what are called Core courses in American universities.
So where students should have been given first and foremost better communications skills, and the ability to work in teams, they were instead given detailed knowledge of science and history, formulas about systems for the former, catalogues for the latter.
Later, when I joined Sabaragamuwa University, having already been involved in developing curricula for the degree courses it was developing, both for new students and for those who had completed AUC Diplomas, I decided I should study Core courses as they were being developed in the United States where they had first started.
They had developed initially because American schools were not like British ones from which students could proceed to specialize because they had been provided there with soft skills and wider knowledge. American High Schools as they were called provided more basic education and those going on to university needed catch up teaching as to the knowledge and skills better schools provide.
But in the early 19th century there was not too much of this. Later, by the end of the 20th century, the few core subjects introduced a century earlier had to be expanded as the range of skills needed for productive employment in a changing world also expanded. Derek Bok, President of Harvard for many years, led seminal changes to the system, which I was able to study. And though we did not do as well as Harvard, I think the Core courses we started here back in 1997 equipped our students well for the world of work. And I was able to introduce something similar at the Military Academy when we looked after their degree course.
But none of this had a wider provenance. So that is why, a quarter of a century after Prof Aluwihare showed the way, we have the Prime Minister in his budget speech highlighting ‘the need to build a knowledge-based economy, and the need to promote sport, particularly among the young generation. The Prime Minister also points out that education reforms are important to tackle issues especially among unemployed young generation’.
I would find this funny, if it were not so tragic. Six years ago, when I wrote to him to say the country needed reforms instead of the elections that had turned into his substitute for action, I drew attention to the need of the following –
a) A new Universities Act that provides meaningful training that promotes employability free to those who need it, whilst facilitating the establishment of other centres of excellence through private/ public partnerships
b) A new Education Act that ensures holistic education, with greater stress on skills and competencies that are developed through extra-curricular activities such as Sports and Social Service and Cultural Activities
But of course nothing happened. And if the inaction of the last year is anything to go by, nothing more will happen and instead we will simply hear more and more platitudes about the need for reform.
This is the sadder in that reform would be so easy. I have no regrets personally about having resigned from the post of State Minister of Higher Education five years ago, in time to avoid all taint of Yahapalanaya crookedness and incompetence. But it was sad for the country since those who took over – and indeed had been put on top of me – had no idea about what was required. The last Chairman of the UGC did try, and I am sure the present one will also try. But what the country needs is thoroughgoing reform that is based on general social needs, and that is inconceivable to those stuck in the ivory tower concept which we still cling to, in terms of not 20th century but rather 19th century British models.
The basic principles that we should work on include –
a) Breadth of knowledge in context, not depth that is essential only for those going on to do research
b) Better communication skills including teaching skills that will facilitate the sharing of knowledge
c) Thinking skills that promote innovation and analysis of different perspectives
d) Social awareness and sensitivity that contributes to coherent and productive planning and action in the world of work
Together with these let me draw attention to something that is particularly relevant to what we are concerned with, the research that students have engaged in. I should note though that when I was responsible for preparing a new curriculum for what was a new university, way back in 1997, I was not keen on what was described as a thesis. Not only did I feel that our undergraduates would simply reproduce material culled from others, I also thought that they would not be properly supervised. And though I agreed in the end to what we called a dissertation, I insisted on an oral viva because it was vital to check that students understood what they were presenting.
This may seem excessively cautious, but those were days in which plagiarization was rife in the universities. I recall interviewing someone for a senior position who had no idea of what anything in his master’s thesis meant. His response was that some authority had said this – which the thesis itself had not acknowledged – and I sent a rocket to the Colombo university Vice-Chancellor who was gracious enough to admit their error. He promised that in the future they would insist on an oral test to check on the understanding of the candidate as to what was in the thesis. But whether this happened I do not know, and of course nothing could be done about those masquerading as scholars on the basis of what they had copied.
The oral test was essential, but I am glad Sabaragamuwa has moved further on this and makes students present their research and respond to questions. But I believe too that we should go further, and ensure that research is practical and at undergraduate level oriented towards community development.
This was one of my last proposals when I was Minister, when I suggested to Vice-Chancellors that they should focus all research at undergraduate level on the area in which they were situated. So political science students could for instance look at the work of Grama Niladhari Divisions and Divisional Secretariats, economists could look at enterprises and employment opportunities in the area, sociologists at family structures and welfare statistics, English students at English education, Sinhala students at writing skills amongst students, geography students at water problems – which six years ago I told the then President was one of the greatest problems we had to face, something that thankfully, if six years too late, the current President has acknowledge.
I suggested then to the Vice-Chancellors, who seemed to think this a good idea though they promptly forgot about it after I resigned, that the students of each final year should focus on one administrative division, so that the university could then prepare a comprehensive development plan for that division based on the research of its students. That research would of course have included consultation of the people of the division, consultation that now rarely occurs when development plans are made.
Politicians will of course think this unwarranted interference, but in a context where many politicians can neither think nor plan, there has to be some sensible input. Of course I may be being unfair to politicians in this area, for I had thought of the need for such interventions after much work in the North and East. In those areas what were termed District and Divisional Development Committees had in the period between 2010 and 2014 been entrusted to scoundrels such as Bathiudeen and Hisbullah and Piyasena – and two of them, enthroned then, went on to engage in even greater destruction over the next five years too.
Such individuals would resent academic involvement but I have no doubt sensible politicians would welcome this. Of course the plans prepared by universities would be subject to further discussion, but they would provide a better basis, with greater factual input, and greater input as to what the people want, than any plans prepared by politicians or even any central government agency could have.
So let me leave you with this suggestion, which I hope the Dean and the Faculty will take up, with the support of your new Vice-Chancellor. Focus on a socially productive outcome in your next cycle of research, one that will allow staff and students of the Faculty who are concerned citizens to work together coherently to make things better for the community in which you are situated. That is what the university of the 21st century should make a priority, development with precise knowledge, deeper understanding, and social sensitivity.
Features
So, who is going to tell the rest of the world?
Series: The greatest digital rethink, Part V of V – Series conclusion
Five instalments. Five levels of education. One recurring pattern: the countries that ran the experiment are retreating, the countries that watched them are still paying the entry price. This final column asks the question the international education community has been carefully avoiding: does anyone actually learn from anyone else, or do we just take turns making the same expensive mistakes?
What five parts told us
Let us briefly take stock. In Part I of this series, we traced the arc of three decades of digital enthusiasm in education, from the early computer labs of the 1990s through the tablet explosion of the 2010s, to the pandemic acceleration and the emerging backlash that defines the present moment. In Part II, we watched Sweden take tablets away from preschoolers who should never have been given them in the first place, and Finland legislate to return the pencil to its rightful place in the primary classroom. In Part III, we confronted the paradox at the heart of secondary school de-digitalisation: governments triumphantly banning the phone in the student’s pocket while quietly expanding the data systems that monitor their every digital interaction. In Part IV, we sat in the university exam hall, a room that had been pronounced redundant 20 years ago, and watched it fill up again with students writing with pens, because the large language models (LLM) like Chat GPT, had made every other form of assessment untrustworthy.
The inconvenient asymmetry
There is a concept in international education research, ‘asymmetric correction’, that describes this phenomenon with academic precision. It means, in plain language, that the systems with enough money, data and institutional capacity to discover that an experiment has gone wrong can afford to correct it. The systems without those resources cannot, and often do not even know the correction is needed until the damage is visible in their own classrooms and their own assessment results.
This is not merely an abstract inequity. It has a specific mechanism. The countries now de-digitalising, Finland, Sweden, Australia, France, the UK, have had 20 or 30 years of experience with school digitalisation. They have run multiple cycles of national assessments. They have PISA data going back decades. They have teacher unions vocal enough to flag classroom deterioration before it becomes a crisis. They have the research infrastructure to connect a policy change to an outcome measure and draw a conclusion. When their scores drop, they investigate. When the investigation points at screens, they act.
The evidence that was always there
One of the more unsettling conclusions of this series is that much of the evidence driving the current de-digitalisation wave was available considerably earlier than the policies it has inspired. The finding that handwritten notes produce better conceptual understanding than typed ones was published in 2014. The OECD’s analysis showing that more computers do not produce better learning outcomes appeared in 2015. UNESCO’s concerns about platform power and datafication in education have been articulated consistently for years. The distraction research, documenting that students with open laptops in lecture halls perform worse, and drag their neighbours down with them, has been accumulating for well over a decade.
None of this stopped the rollout. The tablets arrived in the Swedish preschools. The 1:1 device programmes expanded. The learning management systems embedded themselves. The AI proctoring tools were procured and deployed. Evidence that gave pause was routinely absorbed into a narrative about implementation, the problem was not the technology, it was how it was being used; give us better training, better platforms, better connectivity, and the results will follow. The results, in many cases, did not follow. But by the time that was clear, the infrastructure was in place, the contracts were running, and the political cost of admitting the bet had been wrong was prohibitive.
What changed was not the evidence, it was the political permission to act on it. PISA 2022 delivered declines dramatic enough to be impossible to attribute to anything other than something systemic. UNESCO issued what amounted to an institutional mea culpa. And a sufficient number of teachers, in a sufficient number of countries, were by then willing to say publicly what they had been saying in staffrooms for years: that the screens were not helping, and in many cases were actively in the way.
What a responsible global policy would look like
This series is not a manifesto against technology in education. It has never argued that. Screens are indispensable tools, for accessing information, for enabling collaboration across distance, for serving students whose accessibility needs require digital solutions, for supporting the administrative and logistical complexity of modern educational institutions. The argument is not against technology. It is against the thoughtless, evidence-free, vendor-driven acceleration of technology in contexts where it undermines the very foundations it is supposed to strengthen.
A responsible global education policy would, at minimum, do several things that the current system conspicuously fails to do. It would require that the evidence base for large-scale digital procurement be genuinely independent of the vendors supplying the technology. It would insist that the learning from early-adopter systems, including the learning about what went wrong, be actively communicated to late-adopter systems before, not after, they make the same investments. It would treat the question of appropriate technology use at different ages and in different pedagogical contexts as a matter of ongoing empirical inquiry, not a settled ideological commitment to ‘more is better.’ And it would hold to account the international organisations and development banks that have promoted digital solutions to educational problems without adequate attention to long-term cognitive and social outcomes.
None of this is technically difficult. The knowledge exists. The research is available. The lesson is sitting there in the PISA data, in the Swedish preschool curriculum reversal, in the UK university exam halls filling up with students holding pens. The question is purely one of political will, and of whether the global education community considers it acceptable to keep selling a model it is quietly dismantling at home.
Who decides what technology is for?
Beneath all the policy detail in this series lies a question that is fundamentally political rather than technical: who gets to decide what role technology plays in education, and in whose interest do those decisions get made? The answer, across the period this series has covered, has too often been: vendors, with governments following at a respectful distance and parents and teachers arriving to the conversation after the contract is signed.
De-digitalisation, for all its imperfections, its occasional moral panic, its selective use of evidence and its tendency to become a political signalling exercise, represents something important: a reassertion that educational technology is a means, not an end, and that the people who should determine how much of it to use are educators, researchers and communities, not quarterly earnings reports. The fact that Finland chose to legislate, that Sweden chose to buy books instead of tablets, that Queensland schools now require phones to be away for the day, often collected, or switched off, from the moment students arrive and found their playgrounds transformed, these are acts of pedagogical agency. They are an insistence that schools are for children, not for platforms.
A final word
There is nothing wrong with technology in education. There is something very wrong with the assumption that more technology is always better, and something worse with the global system that allows wealthy nations to learn that lesson expensively, correct it quietly, and then export the uncorrected version to everyone else.
The pencil did not disappear because it failed. It was sidelined because screens arrived with better marketing. It is coming back, in Finnish classrooms, in Swedish preschools, in Australian playgrounds, in university exam halls, not out of nostalgia, but because 30 years of evidence have converged on an uncomfortable truth: some things, it turns out, require your full attention, your physical hand, and the irreplaceable cognitive effort of a human being working without a shortcut.
That is not a retreat. That is a reckoning. And the only question left worth asking is whether the rest of the world will get to benefit from it before they have to discover it for themselves.
SERIES COMPLETE
Part I: From Ed-Tech Enthusiasm to De-Digitalisation | Part II: Phones, Pens & Early Literacy | Part III: Attention, Algorithms & Adolescents | Part IV: Universities, AI & the Handwritten Exam | Part V: Who Is Going to Tell the Rest of the World?
Features
New kid on the block – AI drug prescriber from the US
Artificial intelligence (AI) in healthcare has come to stay and is a well-recognised development over the last decade or so. AI has now progressed on to even the ability to execute quite a few tasks and manoeuvres that were once the sole duties of doctors. Certain AI programmes are now designed to make tricky diagnoses, offer mental counselling, detect drug interactions, read and diagnose images, forecast results, and review scientific articles, to name a few amongst other capabilities. As the aptitudes of AI increase, the roles of doctors are likely to change. In the future, there is a real possibility that physicians would increasingly be placed in supervisory roles in semiautonomous systems, while retaining responsibility but with reduced independence.
Philosopher Walter Benjamin, in the 1930s, wrote that photography and cinema would have a telling effect on paintings and painters. It was argued that the introduction of visual images would render painting and painters quite obsolete. Many belittled the artistic value of photographs, just as today, many ask whether AI can truly understand illness or empathise with discomfort. The opponents of photography theorised that original works of art, such as paintings, had a so-called aura and that there was something special about an original artwork compared to a reproduction as a photo image, and that the painting echoed its singular history and unique trajectory through time, space, and social meaning.
Today’s doctors have something comparable. Their professional authority was grounded in their unique training, the practical wisdom that they had accrued, their face-to-face presence with patients, and their nuanced clinical judgment. Like an original painting, medical expertise appeared singular and inseparable from the clinician who exercised it rather than from the tools or institutions that supported the physician’s practice.
Now enters the latest AI initiative in healthcare. As documented in the Journal of the American Medical Association (JAMA) on the 13th of April 2026, it is the very first AI DRUG PRESCRIBER. It originated in the state of Utah of the United States of America, which is the 45th state admitted to the Union on the 4th of January 1896, and is well-known for its unique geography, including the Great Salt Lake and its “Mighty 5” national parks: Zion, Bryce Canyon, Arches, Capitol Reef, and Canyonlands.
In January 2026, the State of Utah publicised a first-of-its-kind partnership with an AI company to develop an AI-based programme to prescribe medications without physician involvement. The AI prescriber package sold by the company Doctronic is claimed to conduct a “comprehensive medical assessment” that “mirrors the clinical decision-making process a licensed physician would follow“. Originally, it was intended to focus on prescription renewals, and the software is designed to prescribe almost 200 drugs, including corticosteroids, statins, antidepressants, hormones, and anticoagulant agents. It has the potential to develop into an autonomous system that could even provide original prescriptions without the involvement of doctors.
There are perceived advantages to AI prescribing in a world facing shortages of primary care physicians, as well as certain specialists. The public health goal is to make sure that patients have access to safe, effective drugs and continue receiving them for as long as it is appropriate. There are documented scientific studies in Western countries on non-adherence, failure to take the drugs of a first prescription, and failure to get refill prescriptions. True enough, AI could reduce pervasive medication errors, enhance process efficiency, and free physicians to focus on complex diagnostic tasks or human-to-human interactions.
Yet for all that, technology-driven revolutions can also cause damage, create waste, and even destabilise the medical connection. They could reduce the patient-clinician encounters and substantially reduce the prospects for physicians to spot other problems and for patients to raise anxieties and ask questions. Doctors have to go through a rigorous process of training and demonstration of clinical fitness to be allowed to practice medicine. AI prescribers face no equivalent safety process. AI companies generally do not openly reveal the precise operational details of the software’s abilities to make medical decisions. In the Utah deal, generalisations were offered, including that the AI prescriber is “trained on established medical protocols,” and that its algorithm continues to progress through “feedback loops.” However, they are far from the absolute detailed guarantees that training of a physician offers.
In the American System of Governance, most states have long maintained foundational laws for dispensing medicines, positioning licensed physicians and pharmacists as essential caretakers and even as gatekeepers. Federal Law requires that any drug that “is not safe for use except under the supervision of a practitioner licensed by law” must be dispensed only “upon a written prescription of a practitioner licensed by law“. AI prescribers are not licensed “practitioners” of medicine, and here, Utah has waived state requirements. It has waived State Laws for businesses with novel ideas deemed potentially beneficial to consumers.
Under the main FDA statute, an AI prescriber comes under an “instrument, apparatus, implement, or machine clearly intended for use in the cure, mitigation, treatment, or prevention of disease,” which makes it an FDA-regulated medical device. The 21st Century Cures Act of 2016 created exemptions for software involving administrative support, general wellness, or electronic record storage. For clinical software, the FDA has generally exercised enforcement discretion only for tools that aid physician decisions. By design, AI prescribers remove the physician, meaning that FDA oversight is required.
However, in the Utah deal, the company has apparently not attempted to approach the FDA about the technology, thereby working on the presumption that the FDA does not regulate the practice of medicine. True enough, Federal Law and the FDA itself express that the FDA does not regulate the practice of medicine. However, Federal Law also emphasises that medical devices and drugs must be legally sold and used within a legitimate patient-clinician relationship. Federal Law does not permit the replacement of physicians with unlicensed computers.
The scientific aspects of the conundrum imply that the current political administration appears to be disregarding some of the federal oversight. Since its 2025 inauguration, the executive branch of the current administration has rescinded previous AI governance orders, encouraged the removal of policies that might impair innovation, and issued an executive order aimed at reducing federal funds for states that strictly regulate AI. The USA Commissioner of Food and Drugs has clearly emphasised the need for AI innovation. Given this antiregulatory environment for AI, the prospect of federal intervention against initiatives like AI prescribers appears to be quite slim.
As federal and state regulators retreat, private parties have stepped in. The Joint Commission (TJC), a private, non-profit organisation that functions as the primary accrediting body for healthcare organisations, recently released non-binding guidance urging healthcare organisations to establish internal AI governance structures and rigorously measure outcomes. The success of AI prescribers will ultimately depend on the acceptance of health systems, which should demand robust evidence of safety and effectiveness, optimally in the form of clinical trials.
Tort law, a branch of civil law that deals with public wrongs such as situations where one person’s behaviour causes some form of harm or loss to another, remains a potential avenue for addressing patient harm because Utah’s agreement leaves such remedies intact. However, injured patients face significant hurdles. Courts will have to determine whether AI could be held to the same standard of care as a human physician. A product liability lawsuit would typically require a plaintiff to show that there was a reasonable alternative design, a challenge for AI black-box technologies. Furthermore, companies might argue that patients “assumed the risk” of using the AI prescriber. However, that is not a complete defence.
AI prescribing would be safest under concurrent state and federal oversight. Yet Utah has granted a state waiver, and FDA compliance has not been demonstrated. Other companies may take the lesson that they can bypass federal safety standards, and they may race into the market to ensure they are not left behind.
Some examples beg for caution. The FDA fell behind in regulating flavoured e-cigarettes, which are now ubiquitous and have contributed to a youth e-cigarette epidemic, which has even reached Sri Lanka. The sheer scale of the unauthorised market and the subsequent legal tactics used by tobacco companies turned premarket requirements into a mere technicality. If AI prescribing becomes the industry standard before safety and liability frameworks are established, the power problem may render future regulation infeasible.
Although AI offers the promise of increased efficiency and expanded access, the evasion of legal obligations by early movers raises profound concerns. The company that is marketing the AI Prescriber is operating in a unique legal “grey zone” that has sparked intense debate among regulators and medical associations.
Incorporating AI into modern health care must be evidence-based and responsible. Physicians and health systems should insist that AI technologies should not be allowed to bypass long-standing and proven legal guardrails governing medical products. That needs to be the axiom that should apply not only to the Western nations but to the whole wide world.
by Dr B. J. C. Perera
MBBS(Cey), DCH(Cey), DCH(Eng), MD(Paediatrics), MRCP(UK), FRCP(Edin), FRCP(Lond), FRCPCH(UK), FSLCPaed, FCCP, Hony. FRCPCH(UK), Hony. FCGP(SL)
Specialist Consultant Paediatrician and Honorary Senior Fellow, Postgraduate Institute of Medicine, University of Colombo, Sri Lanka.
An Independent Freelance Correspondent.
Features
From the Handbook for Bad Political Appointments
The Geathiswaran Chapter:
Dr. Ganesanathan Geathiswaran, Sri Lanka’s Deputy High Commissioner in Chennai is in hot water, dragging in with him the Foreign Ministry as well as the Sri Lanka government into a worthless controversy. It stands as a classic example of a misplaced political appointment to a sensitive public position paid for by hapless Sri Lankan taxpayers. And that too by a government that came to power promising not to politicise appointments.
Why would a meeting between a Sri Lankan diplomat and a group of fishermen in South India in the last week of March 2026 be controversial? After all, illegal fishing in Sri Lankan waters by South Indian fishermen from the Tamil Nadu area, which negatively impacts the livelihoods of mostly Tamil-speaking Sri Lankan fishing communities, is a perennial problem that neither Sri Lankan nor Indian governments have been able to resolve. This is also a consistent political issue in Tamil Nadu politics. In this context, a Sri Lankan diplomat meeting local fishermen might well be within his job description. But the issue is how and where such a meeting should take place. The bottom line is that it should not be a public event.
Speaking to The Hindu on 5April 2026, Geathiswaran insisted his presence in the meeting was a “routine visit” and that the event was not organised by any political party. He also said, “I’m not here to do politics” and “I have nothing to do with politics.” He further insisted, “I did not take part in any political campaign. It was in an open area along the seashore. The meeting was not on a stage and in a public area.” These utterances show both Geathiswaran’s naivety, woeful lack of experience and understanding of the nature of politics in the region where he is our country’s chief diplomat.
Be that as it may, let us look at the optics and substance of the said event. According to information circulating in the media in both Sri Lanka and India, the Deputy High Commissioner attended a meeting with local fishermen in Puducherry. It was not a closed-door meeting. It appears, the Sri Lankan diplomat was invited to the event or it was coordinated by Jose Charles Martin, the leader of the newly formed political party, Latchiya Jananayaga Katchi (LJK). Though launched only in 2025, the LJK has been making inroads into Tamil Nadu politics mostly funded by the business interests and funds of Martin’s father, the well-known lottery tycoon, Santiago Martin. LJK joined the BJP-led NDA in the ongoing Puducherry Assembly Elections of 2026. Moreover, as indicated in the photographs in circulation, one can easily see the presence of several BJP politicians including V. P. Ramalingam, BJP’s Puducherry president and a candidate in the Raj Bhavan constituency.
Members of Martin’s family are craftily aligned with different Tamil Nadu political formations. Jose Charles Martin himself is contesting the Puducherry electoral area as a BJP ally, while his mother is contesting from the AIADMK, and his brother-in-law is contesting as a candidate of the Tamilaga Vettri Kazhagam (TVK) party.
Therefore, Geathiswaran’s assertion that the event was not organised by a political party is blatantly false. Further, the event does not become non-political just because of the absence of a stage just as much as a stage does not provide political attributes merely because of its higher elevation. It is unacceptable that a diplomat hand-picked by the Sri Lankan President for the important station of Chennai, thereby depriving the appointment of a senior career diplomat with years of work experience and awareness of political nuance and optics, can be allowed to be this naïve.
It is in this context that Pawan Khera, a senior leader of the Indian National Congress, complained in an X post on 4 April tagging the Indian External Affairs Minister noting that Geathiswaran’s participation in the meeting was “a gross violation of the 1961 Vienna Convention on Diplomatic Relations”, according to which “diplomats ‘have a duty not to interfere in the internal affairs of that State.’” He also noted in his post that the diplomat was invited by the leader of the LJK and also referred to the presence of senior BJP politicians. Leaving aside the overemphasis of the Vienna Convention, which in this instance makes no sense, the issue at hand is the complete lack of common sense on the part of the Sri Lankan diplomat that allowed this controversy to arise in the first place. Despite his insistence on not engaging in politics, which in the case is likely true, this was very clearly a political event, politically conceived, perceived and packaged, organised by a political party, and conducted in the presence of allied politicians who were contesting in a local election. As a foreign diplomatic representative, Geathiswaran should have the cerebral wherewithal to make the distinction or at least seek guidance from his superiors at the Foreign Ministry in Colombo.
Diplomats need not shy away from controversy if it makes sense and benefits the nation. But the incident under reference is purely nonsensical from any perspective. This brings me back to Geathiswaran’s appointment as Sri Lanka’s Deputy High Commissioner in Chennai, itself. What unique experiences did he bring to the post? Of course, he is Tamil-speaking. So are hundreds of thousands of other citizens in the country including potentially competent, well-trained, intelligent and experienced career diplomats. I am not saying that political appointments are necessarily unfavourable, though not ideal unless they bring to the service expertise that the Foreign Service does not have. But what quality and qualification does Geathiswaran possess for the position that is lacking in a career foreign service officer?
Does he bring in access to the different segments of Tamil Nadu political landscape that no one else has? If so, should this controversy not have arisen in the first place, owing to the good connections to the entire political spectrum? In short, he brings absolutely nothing to his office and the country he represents. He also does not have any diplomatic or any other public or private sector experience that would have injected sense and nuance into the present posting. His only qualification is the close political connection to the NPP through family.
This fiasco brings to mind some ideas I presented in 2024 in the government’s own newspaper, the Observer two weeks before the NPP government was established and about one month after President Dissanayake assumed office. Since those conditions still remain valid and the present incident raises the same alarm I raised then, I think it is worth reflecting on them yet again:
“During the last three decades, particularly during the Rajapaksa administration, Sri Lanka’s Foreign Service saw a significant nosedive … In real terms what this means is, the Foreign Service has been encroached by individuals purely based on their political and nepotistic connections, with little or no regard for requisite qualifications, expertise or experience. This is observed not only at ambassadorial level, but also right down to the junior levels in our overseas missions … The main reason for the sorry state of the Sri Lanka Foreign Service is that it has been problematically and parochially politicised over a long period of time, without any pushback … Political appointments are a serious problem. Due to the appointment of completely unqualified individuals on political patronage, there are very few intelligent and well-trained personnel in our embassies in the major cities of the world who are able to proactively work in the country’s interest, when problems arise at the global level. Furthermore, it is also not apparent if there are officials in the Ministry who can advise their unenlightened political superiors without fear and stand their ground on principle. This situation has come about as a matter of simple personal survival and bread-and-butter purposes, owing to which both the larger interest of the Service and self-respect of officers have been clearly compromised.”
Is this not what the Chennai incident also indicates? Geathiswaran being a wrongful appointment is one matter. But it also appears that he did not even have the common sense to seek advice before the meeting in Puducherry or such advice was simply not forthcoming or heeded, as political appointees are generally considered a know-it-all bunch who have the ears of the political hierarchy, and therefore above the norms and regulations that apply to mere career officials.
For many of us the advent of the NPP to power signified the dismantling of the culture of political patronage in which diplomatic postings were rewards for loyalty and friendships. It took less time for the present government than others to go against its own repeatedly stated pre-election positions and to stuff the Foreign Service with incompetent individuals. The present fiasco authored by one of these appointees exemplifies the consequences of this continuing malpractice.
Let me leave readers and government apologists with the words of Tom Nichols, former professor at the U.S. Naval War College about Trumpian ambassadorial appointments, as this applies to our country too: “[With some of his ambassador choices], Trump has elevated diplomatic incompetence to an art.”
Sri Lanka just might outdo the mighty US President on this score.
-
News6 days agoCEB orders temporary shutdown of large rooftop solar systems
-
Features6 days agoFrom Royal College Platoon to National Cadet Corps: 145 years of discipline, leadership, and modern challenges
-
Latest News5 days agoPNS TAIMUR & ASLAT arrive in Colombo
-
News7 days agoAnura Solomons passes away
-
Features6 days agoCIA’s hidden weapon in Iran
-
Latest News5 days agoPrasidh, Buttler set up comfortable win for Gujarat Titans
-
News2 days agoPNS TAIMUR & ASLAT set sail from Colombo
-
Business7 days agoSL’s economic outlook for 2026 being shaped by M-E conflict
