Comments (2)
Hello,
I've come across another CGL benchmark, BeGin (https://arxiv.org/pdf/2211.14568.pdf), in table 5.c. the performance of aromaticity-CL is reported, they have a result of 0.286, while in CGLB, this is ~78%. I would expect the result to be more similar.
Kr
Hi,
This seems to be a new benchmark and I haven't checked its code. But I think such a huge difference should be caused by the number of classes included.
As explained in the Appendix 1.1 of our paper, 'For the G-CGL datasets, classes removed from Aromaticity-CL are {2,
3, 4, 8, 35, 36, 37, 38, 39, 40, 41} since they contain less than 20 examples and are causing difficulties
for model training. The other 30 classes of Aromaticity-CL are kept and constructed as 15 tasks.'
We found that these very small classes are causing difficulty for the model to learn, and even the joint train cannot perform well. After checking, the difficulty is that the model is overfitted to the very few training data, therefore is not performing well on the test set of these small classes. This difficulty is actually concerned with how to design the GNNs with better generalization power, and is not the focus of continual learning. If the model cannot learn well on each task, then it does not make sense to further discuss the forgetting problem, which is the focus of continual graph learning. Therefore, we chose to only keep part of the original datasets to ensure that the constructed tasks are suitable for evaluating continual learning models.
from cglb.
Thank you for the clarification, it can indeed be the case. It is a interesting question to think about when continual learning performance is harmed due to external factors such as overfitting.
from cglb.
Related Issues (20)
- Redundant training in testing phase? HOT 2
- Performance matrix visualization for GCGL is not working correctly HOT 2
- add more units to the output layer in LWF HOT 2
- Problem in utils.py HOT 2
- Reddit Dataset Batch Size Issue HOT 2
- GEM baseline in class-IL for GCGL has wrong indentation for optimizer step? HOT 2
- GCGL set_random seed is never called HOT 2
- Problem in function pipeline_task_IL_inter_edge_minibatch HOT 1
- GEM Memory_data HOT 1
- GCGL, TWP error when using higher GCN hidden units HOT 1
- Question about hyper-parameter search in GCGL HOT 2
- SIDER-tIL Jointtrain Cannot be reproduced HOT 1
- Class Incremental for Node Classification EWC and MAS cannot be reproduced HOT 2
- Error occurred for jointtrain in GCGL HOT 3
- error report when implementing GEM on Tox21 datasets HOT 1
- ValueError from GEM on Tox21 HOT 1
- Is ```observe_class_IL_batch``` function considering inter-edge connections in ```pipeline_class_IL_no_inter_edge_minibatch```?
- Environment configuration
- non-continuous categories
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from cglb.