Recent Research by CRT in AI PhDs
The full articles can be accessed by clicking on authors’ names.
M. Adeel Haffeez
Title: Depth Estimation Using Weighted-Loss and Transfer Learning
Summary: I present a method that improves depth estimation accuracy from 2D images. I proposed a novel loss function that combines Mean Absolute Error (MAE), Edge Loss, and Structural Similarity Index (SSIM) for enhanced robustness and generalization. I evaluated several pre-trained DenseNet and EfficientNet encoder-decoder models on the NYU Depth Dataset v2. The EfficientNet architecture, paired with a simple decoder and the optimized loss function, delivered outstanding results, achieving RMSE, REL, and log10 scores of 0.386, 0.113, and 0.049, respectively, outperforming existing models in accuracy and robustness.
Yumnah Hasan
Title: Interpretable Solutions for Breast CancerDiagnosis with Grammatical Evolution and Data Augmentation
Summary: Medical imaging diagnosis increasingly relies on Machine Learning (ML) models, but is often hindered by imbalanced datasets and limited interpretability. This paper proposes using the STEM synthetic data generation technique with Grammatical Evolution (GE) models, known for their interpretability. STEM combines SMOTE, ENN, and Mixup to address imbalance issues. Testing on DDSM and WBC datasets against standard ML classifiers, GE-derived models show the highest AUC while remaining interpretable. This highlights the effectiveness of STEM in producing interpretable ML models for medical imaging diagnosis, addressing both imbalance and interpretability challenges in the field.
Additional Notes: This paper was nominated for the best paper award in Evostar 2024. It also won the best poster award. I received an outstanding student nomination for this award.
Muhammad Asad
Title: Beyond the Known: Adversarial Autoencoders in Novelty Detection
Summary: In novelty detection, the aim is to classify new data points as inliers or outliers based on a training dataset of inliers. We employ a lightweight deep network, leveraging a deep encoder-decoder framework to generate a reconstruction error. This error is used to calculate a probabilistic novelty score. Our research introduces two innovations: first, we linearize the manifold representing the inlier distribution to interpret the novelty probability in relation to manifold’s local tangent space coordinates. Second, we enhance the network’s training protocol. Our methodology demonstrates superior performance in identifying the target class compared to recent advanced methods across multiple benchmark datasets.
Kevlyn Kadamala
Title: Enhancing HVAC control systems through transfer learning with deep reinforcement learning agents
Summary: Traditional HVAC control systems relied on rule-based schedulers, but deep reinforcement learning offers a data-driven approach without explicit programming. However, learning effective policies from scratch can be time-consuming. Transfer learning with pre-trained models can save time and resources by leveraging existing knowledge. In this study, we use reinforcement learning to pre-train and fine-tune neural networks for HVAC control. An RL agent was trained in a building simulation environment and then fine-tuned on different environments simulating varied weather conditions and buildings. Results showed transfer learning agents outperformed rule-based controllers, with improvements ranging from 1% to 4% compared to agents trained from scratch.
Updates in AI Policies and Regulations
Interim Guidelines for Use of AI in the Public Services (updated February 2024)
AI - Here for Good: Progress Report on the National AI Strategy
Ministers welcome agreement on EU Artificial Intelligence (AI) Act
A risk-based approach: The Regulatory Framework defines 4 levels of risk for AI systems
AI innovation package to support Artificial Intelligence startups and SMEs
CRT in AI Students Visited The University of Amsterdam
The CRT in AI Programme was thrilled to bring 40 students to The University of Amsterdam in April 2024 to hear from AI experts from The Amsterdam AI group. Led by Paul Buitelaar, Derek Bridge and Suzanne Little, this 3 day visit, included a cohort build exercise at the Rijksmuseum and free time to network with one another. Overall, the first ever international CRT in AI cohort visit was a positive one, we are looking forward to exposing our CRT in AI researchers to more learning opportunities in the near future!
Anonymous feedback:
Amsterdam AI trip was a very enriching team-building experience, talking to PhD students from Irish universities about their research and learning about the country together. Also discussion about various facets of AI research happening in Netherlands was also quite interesting. Overall very good experience.
4/5 « A good few days in Amsterdam; fun meeting up with the different cohorts and leaving with some perfume from the 16th century. The sandwiches provided along with the opportunities (notably postdoc) were good!
4.5/5 Amsterdam AI trip was a lovely experience. From the cohort building activities to the free time to allow us explore and the topics for the presentation. The presentations were informative and the entire program ran smooth. It was definitely worth my while
Deep Learning: Challenges faced by non-EU PhD researchers applying for Travel Visas
By: Janet Choi Sr. Research Co-ordinator, SFI CRT in Artificial Intelligence Programme Manager
In June of 2023, the team at CRT in AI co-ordinated a visit to The University of Amsterdam scheduled for April 2024. For our non-EU PhD Researchers, this meant travel visas were required, and I wanted to help secure these appointments for them. I only needed 23, so I resolved to the fact that it would be a simple phone call.
“Just make the call to VFS Global in Dublin, ask for the Embassy of the Netherlands, request for a group appointment.’’ I said to myself.
Making the call, I was placed on hold, waiting for the standard opening welcome from the service agent, who couldn’t manage my request. I was transferred to a case manager. Who I thought was really going to help me finally. Getting rather excited, I thought I would get my 23 appointments before April 2024. This was June 2023, so loads of time at hand. I was so wrong.
When Reality Meets Expectations
I had a brief call with the case manager, who assured me that the Dublin office would handle all 23 requests in one day. They would email me the details.
I waited for the promised email, which came promptly after the call, an acknowledgement email thanking me for my phone call and dealing with VFS global each time.
I waited, and waited for appointment details. Anticipating details of the times and dates for the in person appointments, my email reminders to VFS mostly unanswered, became commonplace, and days turned to weeks.
Another phone call, attempted again to reach the same case manager, who was so happy to facilitate me initially on the first phone call. Resulted in more than 20+ phone calls by landline and mobile to gain clarity on the appointment dates and times. Between calls, I would receive email acknowledgements. Emails that were digitally signed with a Dublin address, but in actuality the call centre was based in Mumbai.
On several occasions, I received multiple missed calls from unknown numbers, followed by emails from VFS Global stating that they tried to contact me unsuccessfully which delayed the process by 24-48 hours. A satisfaction survey by email would promptly follow the initial email. These numbers were international numbers that I did not recognise, so promptly did not answer these.
A long story short, from the promise of 23 appointments being managed in one day, the conversations each time went on a vertical descent, with appointments reduced, down from all, to 6 per day, spiralling down from 6 to 4 per day and finally to 2 per day.
Separately, calls to help a PhD’s family appointment for 4 (which was requested to be processed together) was split to two individual appointments on two separate days, with the group travelling outside of Dublin on two separate occasions.
This trip, the programme’s first international one, was too important to fail. All the effort spent on meetings, discussions, and countless emails with Amsterdam AI and the PhDs’ time was wasted. The time invested in securing flights and accommodation before applications even felt at risk. I felt sick with worry on multiple occasions, seeing in my mind the students’ disappointment.
Establishing Required Documents for Each Applicant
The Embassy required the following documents, and in future I would advise all the students to do this, have a portfolio ready of all the documents required as listed below on their person when granted a visit to VFS Global. The list below is mandatory.
- Agenda/Itinerary for the visit on visiting site Letterhead
- Letter of Invitation from the institution/conference/event
- Flights and accommodation booked - Travel Itinerary
- Letter of Support from their institution - Sponsor Letter
- Travel Insurance (personal and institutional)
- 3 months Bank Statements
- 3 months’ pay slips
- Fully completed VFS Application
- VFS Global email invitation
- Fee
Lessons Learned
In all, 40 of us made the visit across to The Netherland’s and the visa applications approved by The Netherland’s, springboarded some PhD Researchers to be able to apply for a Schengen Visa.
Considering the monopoly that VFS global has as the single operator for a majority of Embassies who rely on their service, I would say that their service is severely restrictive, and mediocre at best.
From their webpage, they state ‘’leveraging a robust experience in visa application processing, providing governments with a holistic administrative solution to processing passport applications and provision of efficient consular services’’. Which left me with no comment. See vfsglobal.com.
To add, one needs to account for the extra time and funds required for the trip if the appointment is issued for early morning and travel is required outside the Dublin district.
Try to obtain an online appointment first, but my advice is to contact VFS directly, keeping in mind that the applicant will need to persistently seek updates and clarification until a confirmed appointment time and date is provided.
Travelling without the need for a visa is an honour, shedding light on the stark differences in global mobility. This undertaking, underscored the inherent privilege enjoyed by those who can travel freely and frequently, unburdened by the often cumbersome visa application process.
Our Students in 30 under 30 Business Post
Three SFI CRT in AI PhD Researchers have earned recognition in Business Post- newspaper’s annual 30 under 30 list published on Saturday April 27th 2024. The prestigious honour puts a spotlight on Ireland’s promising tech talent. Congratulations to Nicola Rossberg, (supervised by Dr Andrea Visentin and Professor Barry O’Sullivan) Akasha Shafiq (supervised by Dr Paolo Palmieri) and Josh McGiff, CRT in AI Student based at the University of Limerick supervised by Nik Nikolov. These are our bright young stars in Ireland’s Tech scene and the ones to watch!
Akasha Shafiq
Cybersecurity researcher and founder of CyEd – University College Cork, School of Computer Science and Information Technology The SFI Centre for Research Training in Artificial Intelligence – 1st Year PhD.
With a master’s in computer science behind her, Shafiq is now pursuing a PhD through research focused on cybersecurity and privacy-preserving environments. This research could contribute to the security of the internet of things and, meanwhile, Shafiq is also developing an edtech platform to make the world more cybersecurity savvy. Targeting younger users, CyEd gamifies the learning process to teach users fundamental concepts that will help keep them safe online.
Nicola Rossberg
AI Researcher and Community Leader – University College Cork, School of Computer Science and Information Technology The SFI Centre for Research Training in Artificial Intelligence – 1st Year PhD.
Currently, Nicola is conducting AI Research as a PhD candidate at the CRT in AI at University College Cork, Rossberg is examining natural language processing and deep learning – the kind of technologies that underpin LLM like ChatGPT, her research could lead to improved explainability of these technologies – something that is much needed in the industry. Outside of her PhD, she was involved in organising a record-setting edition of the Irish Collegiate Programming Competition this year, where students were challenged to their problem solving, numeracy and computer programming skills.
Josh McGiff
Founder, Educator and Community Leader AI Researcher at The SFI Centre for Research Training in Artificial Intelligence – 1st Year PhD University of Limerick.
Limerick-based McGiff lectures in the University of Limerick’s Immersive Software Engineering course while also completing his own PhD in AI. He is also the founder of the Twitch Ireland community, an online collective of viewers and creators on the popular streaming platform and has developed tools to support thousands of streamers through his business, Streamably. He knows their needs well, having dabbled in streaming himself, helping to popularise the Irish language as he’s done so.
Turing Point Podcast
Episode 4: People Behind The PhD interview with Nicola Rossberg moderated by Josh McGiff. In this interview, Nicola Rossberg, a first-year PhD student with Science Foundation Ireland’s Centre for Research Training in Artificial Intelligence, discusses her work on explainable AI for medical applications, her unconventional path to AI through psychology and data science, and her efforts to tailor explainability techniques to complex medical data, highlighting the challenges and potential of integrating AI into medical practice amidst evolving EU guidelines.
People Behind The PhD w/ Nicola Rossberg
By Josh McGiff
This interview has been taken and adapted from our recorded interview for clarity and brevity.
Hello Nicola, thank you so much for joining me for People Behind The PhD! It’s a pleasure to have you on this episode, I suppose, if we are calling it that (laughs). My main objective is to find out more about the people behind the PhD research within the CRT in AI program.
Tell us about yourself! What does your research entail?
Right, so I am a first-year PhD, I’m actually almost done my first year now, and I’m working specifically in explainable AI, and currently looking at some medical applications of explainable AI and trying to tailor-make explainability methods for medical applications that actually take into account the type of data we’re working with, rather than just being generically applied to a model.
Wow, that’s very interesting. So within the medical sphere is kind of where you’re focusing on. I don’t have much of a biology background, so I feel like jumping into that from an AI point of view must be challenging, right?
Oh yeah, no, it was definitely really tough in the beginning, but I’m really lucky that I get to work with Tyndall National Institute, who are actually here in Cork as well, who are providing all the domain knowledge. So whenever I have any stupid questions where I’m like: “I don’t understand how this works, please help me”. They’re always there to kind of offer support, and they’ve been really great to provide any kind of factors that should be considered when I’m doing my research and when I’m designing my algorithms.
How did you find AI as kind of a topic that you wanted to focus your research on? What made you pick AI as a research area?
So my way to AI was actually kind of weird because my undergraduate was in psychology. So I actually majored in statistics and methodology, and I really liked the math aspect of all of it. I ended up taking a programming class as one of my electives during my final year, and I really liked programming. So I ended up actually choosing data science for my master’s program. During my data science master’s, I then took a deep learning class, which I thought was incredible.
What is your main research question or main objective at the moment?
I’m currently at a bit of a turning point because I finished my first paper. We’re currently in the process of submitting to a few different conferences. My first paper was an application paper, so I was just deploying some different existing explainability techniques to this medical data, seeing what worked, testing out where were the pitfalls of deploying these explainability techniques. Now I’m currently in the process of moving on from there and actually starting to develop my own explainability technique, which is more tailor-made to the medical data. I’m currently writing a lot of code and doing a lot of trial and error to see where there’s possible mistakes and where there’s possible challenges that we could address, and also just doing a lot of reading to find gaps in the research that exists at the moment.
That’s so exciting and impressive about the application paper. So your next step is to dive deeper into your own explainability technique?
Yeah, definitely. One of the problems we’re having with our data that we’re working with is that it’s really multidimensional and multicollinear, which causes a lot of issues with explainability because that means explainability often from one run to another won’t be the same. It’ll change constantly due to the data being so complex. So we’re trying to find different ways of tackling that issue and seeing if we can just make it reliable in a sense. Also because that way it’ll be easier for practitioners to understand how the AI algorithm works and it’ll be easier to actually convince practitioners to use the AI algorithm.
That’s very interesting because I know from a bit of research that I’ve done that some of the issues with getting practitioners and medical people to actually adopt some of the very exciting AI models is because they are not convinced of the methods used for them or they don’t understand what’s going on within them. They would rather use their own traditional methods as opposed to using these new tools just because they simply don’t understand them. Maybe with your kind of work more practitioners can be convinced that these technologies are actually a good thing.
Yeah, hopefully that would be fantastic. I’d be really happy with that. Also with the new AI guidelines coming in from the European Union, we’re kind of looking at more of explainability in AI anyway because it’s necessary for this type of trustworthiness, being able to audit models. So all in all it’s an area that’s really changing a lot right now and it’s really interesting.
In terms of the technologies that you’re using, are you using a neural network approach? Are you using traditional AI? What are your technical approaches?
So for my initial application paper I used a lot of neural networks and it was actually really interesting because initially my type of contribution was just going to be the explainability for neural networks. That was my kind of idea for the PhD, but we kind of ran into a wall in the sense that our more basic machine learning models like random forests were outperforming our neural networks. We didn’t have much of a justification anymore to actually deploy the neural networks because why would you deploy something so complex and use a much simpler model? So now I’m actually looking at some explainability techniques for more traditional models like random forests, decision trees and seeing how they can be adapted to my data type.
That’s very interesting. Do you think you’ll ever try and expand on the neural network side of things again or have you parked that?
I would definitely like to just because I think it’s incredibly interesting and I really like working with deep learning models because it’s also what initially sparked my interest in AI. But I will need a good justification for it before I actually start with it again. [Neural networks are] very different as well and it is so hard to get data in the medical domain.
There’s a lot of discussion at the moment around artificial general intelligence and I’ve looked into some articles about people that believe that we have artificial general intelligence coming quite soon. Some people say it’s within the next decade or so, while there’s also a smallish minority of people that say that it’s never going to happen. I’m curious to hear what you think. Do you think that we are going towards artificial general intelligence?
Well I actually argued with my housemate about this a couple of weeks ago so it’s fine to bring it up (laughs). I think it’s definitely naive to say it’s not going to happen because we thought ChatGPT and large language models weren’t going to happen until a year ago and they came within months and they were just incredibly high functioning. So I think it’s definitely going to happen and with the onset of quantum computing we have no idea what’s actually going to be possible so I’m not going to put a time frame on it but I’d say it’s definitely possible. It’s just a matter of when.
Nicola Rossberg is currently a first year PhD student with Science Foundation Ireland’s Centre for Research Training in Artificial Intelligence.
CRT in AI Upcoming Events 2024 - 2025
Future Skills Showcase 2024
In February 2023, The Future Skills Professional Showcase was held as a two day in person event. The core function was to help actively guide PhD Researchers to understand the skills required from industry professionals to facilitate their development of essential skills for post-graduation career success.
Our programme emphasises on the importance of workplace readiness for PhD researchers, having generated the original concept for the Future Professional Skills Showcase, an agenda focusing on the importance of preparing PhDs for workforce preparation, and employability. Investing in student development, we connected over 150+ PhD researchers with industry partners, promoting research communication and career-readiness.