Comments (6)
Training GAN could sometimes be unstable. Would you mind to visualize some images from the generator to see whether it is good or not? If good, I think 0.15 of D-A should be fine.
from landmark-detection.
Training GAN could sometimes be unstable. Would you mind to visualize some images from the generator to see whether it is good or not? If good, I think 0.15 of D-A should be fine.
Here is some images from generator after 34 epoch:
I find the D-A and D-B value is not reduce from the beginning! and the pictures generated by GAN are quite bad.It seems like the net doesn't learn anything from the dataset! Below is my running log.what's wrong?
BTW,I noticed that when use 300W_Base_Cluster.sh and 300W_Cluster.sh to cluster datasets,their results would be totally different.For cluster-00-03.lst,former will have 16% percentage while latter only have 1% percentage! Is this normal?
seed-4512-26-Aug-at-07-49-40.log
from landmark-detection.
Would you mind to let know how do you visualize these generated images?
Yes, 300W_Base_Cluster.sh' and
300W_Cluster.sh' use different hyper-parameters. I suggest use `300W_Cluster.sh', which one do you use? This is my log for the cluster scrip: https://drive.google.com/open?id=1DtUg0LoZXNuDxeqRtcRxrA4tjrN9rpJa
You can also check my log for the loss of D at https://drive.google.com/open?id=1SZVJHl8tM0G5MOFQmCrx5mB6vFj-aAFu , which decreased to 0.001 finally.
from landmark-detection.
Would you mind to let know how do you visualize these generated images?
Yes,300W_Base_Cluster.sh' and
300W_Cluster.sh' use different hyper-parameters. I suggest use `300W_Cluster.sh', which one do you use? This is my log for the cluster scrip: https://drive.google.com/open?id=1DtUg0LoZXNuDxeqRtcRxrA4tjrN9rpJa
You can also check my log for the loss of D at https://drive.google.com/open?id=1SZVJHl8tM0G5MOFQmCrx5mB6vFj-aAFu , which decreased to 0.001 finally.
Yes.I visualize the images with the visual_freq=5000 option on.I use 300W_Cluster.sh to cluster your style datasets,below is my result after 110 epoch:
seed-4538-26-Aug-at-08-48-45.log
And below is my cluster script and train script:
echo script name: $0
echo $# arguments
if [ "$#" -ne 3 ] ;then
echo "Input illegal number of parameters " $#
echo "Need 3 parameters for gpu devices and detector and the number of cluster"
exit 1
fi
gpus=$1
cluster=$3
batch_size=16
height=224
width=224
dataset_name=300W_$2
CUDA_VISIBLE_DEVICES=${gpus} python cluster.py \
--style_train_root ./cache_data/cache/300W \
--train_list ./cache_data/lists/300W/Original/300w.train.$2 \
./cache_data/lists/300W/Original/300w.test.full.$2 \
--learning_rate 0.01 --epochs 2 \
--save_path ./snapshots/CLUSTER-${dataset_name}-${cluster} \
--num_pts 68 --pre_crop_expand 0.2 \
--dataset_name ${dataset_name} \
--scale_min 1 --scale_max 1 --scale_eval 1 --eval_batch ${batch_size} --batch_size ${batch_size} \
--crop_height ${height} --crop_width ${width} --crop_perturb_max 30 \
--sigma 3 --print_freq 50 --print_freq_eval 100 --pretrain \
--evaluation --heatmap_type gaussian --argmax_size 3 --n_clusters ${cluster}
echo script name: $0
echo $# arguments
if [ "$#" -ne 2 ] ;then
echo "Input illegal number of parameters " $#
echo "Need 3 parameters for gpu devices and detector and sigma"
exit 1
fi
gpus=$1
model=itn_cpm
epochs=50
stages=3
batch_size=8
GPUS=1
sigma=4
height=128
width=128
dataset_name=300W_$2
CUDA_VISIBLE_DEVICES=${gpus} python san_main.py \
--train_list ./cache_data/lists/300W/Original/300w.train.$2 \
--eval_lists ./cache_data/lists/300W/Original/300w.test.common.$2 \
./cache_data/lists/300W/Original/300w.test.challenge.$2 \
./cache_data/lists/300W/Original/300w.test.full.$2 \
--cycle_a_lists ./snapshots/CLUSTER-300W_$2-3/cluster-01-03.lst \
--cycle_b_lists ./snapshots/CLUSTER-300W_$2-3/cluster-02-03.lst \
--num_pts 68 --pre_crop_expand 0.2 \
--arch ${model} --cpm_stage ${stages} \
--save_path ./snapshots/SAN_${dataset_name}_${model}_${stages}_${epochs}_sigma${sigma}_${height}x${width}x8 \
--learning_rate 0.00005 --decay 0.0005 --batch_size ${batch_size} --workers 20 --gpu_ids 0 \
--epochs ${epochs} --schedule 30 35 40 45 --gammas 0.5 0.5 0.5 0.5 \
--dataset_name ${dataset_name} \
--scale_min 1 --scale_max 1 --scale_eval 1 --eval_batch ${batch_size} \
--crop_height ${height} --crop_width ${width} --crop_perturb_max 30 \
--sigma ${sigma} --print_freq 50 --visual_freq=5000 --print_freq_eval 100 --pretrain \
--evaluation --heatmap_type gaussian --argmax_size 3 \
--epoch_count 1 --niter 100 --niter_decay 100 --identity 0.1 \
--cycle_batchSize 8
Seems like the net didn't learn anything!would somethings wrong on the train procedure?
look forward to your kind reply!
from landmark-detection.
I'm not sure, the loss values in your log are quite different from my log.
from landmark-detection.
Ok,i will continue working on it,thanks!
from landmark-detection.
Related Issues (20)
- Pretrained Model of SRT? HOT 3
- How to divide 300VW-Dataset into 3 categories HOT 1
- lk_target_loss return None, scale of batch_locs? How to set forward_max and fb_thresh? HOT 3
- [SAN/SBR] Problems in point_meta about apply_rotate . HOT 1
- Can someone share the paper SRT? HOT 1
- why dont normalized the network outputs to (0-1)? HOT 6
- Problem about training SBR with 300VW dataset HOT 1
- SRT inference HOT 8
- SBR using for our videos
- training failed in SRT heatmaps? HOT 1
- SRT inference get wrong results HOT 6
- SRT results HOT 1
- SBR inference on new video
- how to get arhuments of --face
- SRT inference on video HOT 2
- [Dataset issues] How to download "Synthetic-Face" dataset? HOT 2
- [feature request] SRT improve video inference
- [feature request] SRT inference on multiview images and videos
- ModuleNotFoundError: No module named 'xvision' HOT 3
- Confuse about "batch_interpolate_flow" function in SRT/lib/sbr/interpolate.py
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from landmark-detection.