Comments (9)
do you use pretrained vit? the results of paper are obtained using torchssl, it trained wide resnet from scratch.
from semi-supervised-learning.
error rate is 38%
Which config file are you using?
from semi-supervised-learning.
This issue is stale because it has been open 30 days with no activity. Remove stale label or comment or this will be closed in 5 days.
from semi-supervised-learning.
Hi, can you check the hyper-parameters in this file: https://drive.google.com/drive/folders/1oON5Vyjvb3vWxOQl7hdUl-eh0K-TEPPS.
The config file is slightly different from what we use for reporting the results.
from semi-supervised-learning.
Thanks for your response. I have checked the hyper-paprameter in my log. I found that clip_thresh
was set to False in my log, which do not contain this arguments. However, I donot find clip_thresh
in the config file. The following is my log
[2023-11-24 00:13:36,410 INFO] Use GPU: 0 for training
[2023-11-24 00:13:38,240 INFO] unlabeled data number: 50000, labeled data number 400
[2023-11-24 00:13:39,201 INFO] Create train and test data loaders
[2023-11-24 00:13:39,202 INFO] [!] data loader keys: dict_keys(['train_lb', 'train_ulb', 'eval'])
[2023-11-24 00:13:40,203 INFO] Create optimizer and scheduler
[2023-11-24 00:13:40,211 INFO] Number of Trainable Params: 21436900
[2023-11-24 00:13:40,338 INFO] Arguments: Namespace(T=0.5, algorithm='freematch', amp=False, batch_size=8, c='config/usb_cv/freematch/freematch_cifar100_400_0.yaml', clip=0.0, clip_grad=0, clip_thresh=False, crop_ratio=0.875, data_dir='./data', dataset='cifar100', dist_backend='nccl', dist_url='tcp://127.0.0.1:26868', distributed=True, ema_m=0.0, ema_p=0.999, ent_loss_ratio=0.001, epoch=200, eval_batch_size=16, gpu=0, hard_label=True, imb_algorithm=None, img_size=32, include_lb_to_ulb=True, layer_decay=0.5, lb_dest_len=400, lb_imb_ratio=1, load_path='./saved_models/usb_cv//freematch_cifar100_400_0/latest_model.pth', lr=0.0005, max_length=512, max_length_seconds=4.0, momentum=0.9, multiprocessing_distributed=True, net='vit_small_patch2_32', net_from_name=False, num_classes=100, num_eval_iter=2048, num_labels=400, num_log_iter=256, num_train_iter=204800, num_warmup_iter=0, num_workers=4, optim='AdamW', overwrite=True, pretrain_path='https://github.com/microsoft/Semi-supervised-learning/releases/download/v.0.0.0/vit_small_patch2_32_mlp_im_1k_32.pth', rank=0, resume=True, sample_rate=16000, save_dir='./saved_models/usb_cv/', save_name='freematch_cifar100_400_0', seed=0, train_sampler='RandomSampler', ulb_dest_len=50000, ulb_imb_ratio=1, ulb_loss_ratio=1.0, ulb_num_labels=None, uratio=1, use_aim=False, use_cat=True, use_pretrain=True, use_quantile=False, use_tensorboard=True, use_wandb=False, weight_decay=0.0005, world_size=1)
[2023-11-24 00:13:40,339 INFO] Resume load path ./saved_models/usb_cv//freematch_cifar100_400_0/latest_model.pth does not exist
[2023-11-24 00:13:40,339 INFO] Model training
[2023-11-24 00:22:55,180 INFO] 256 iteration USE_EMA: False, train/sup_loss: 1.8427, train/unsup_loss: 1.5923, train/total_loss: 3.4083, train/util_ratio: 1.0000, train/run_time: 0.0531, lr: 0.0005, train/prefetch_time: 0.0025
What's more, how many GPUs were used during the training process for this log? Will the number of GPUs have a significant impact on performance?
from semi-supervised-learning.
my result is 82.08,same as you ,but the results in paper the error rate is 36% the gap is too large what is the problem.
from semi-supervised-learning.
error rate is 38%
from semi-supervised-learning.
已经解决了
from semi-supervised-learning.
close as no one continues discussing, will re-open if it still has issues.
from semi-supervised-learning.
Related Issues (20)
- FreeMatch SAF loss is a negative value? HOT 4
- Question about the Augmentation on Two-Moon dataset used in FreeMatch and SoftMatch HOT 3
- Bug when running the instance code HOT 2
- how to use timm.model as net HOT 2
- Customize datasets tutorial cant work HOT 2
- I can't load model correctly HOT 1
- colab code can not run in Custom_Dataset.ipynb” HOT 1
- 为什么我在自己的数据集上面,100-600-1200不同的有标签数量训练之后,在测试集的效果是一样的差。 HOT 2
- SAT.ass HOT 1
- About config, how to decide the hyperparameters? HOT 3
- There seems to be something strange when the data is loading HOT 6
- Issues related to voice datasets HOT 1
- R..net..\..m..;M// HOT 1
- How to decide the number of labels in experiments HOT 1
- Questions about batch normalization handling. HOT 1
- Can this run a multilabel problem HOT 1
- Testing models on audio datasets HOT 1
- semilearn get config from file
- Semi-supervised-learning/semilearn/datasets/cv_datasets /cifar.py代码是否有误? HOT 1
- som question about get_cosine_schedule_with_warmup HOT 3
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from semi-supervised-learning.