rtdetr_config.py 2.0 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869
  1. # yolo-free config
  2. rtdetr_cfg = {
  3. # P5
  4. 'rtdetr_n': {
  5. # ---------------- Model config ----------------
  6. ## ------- Image Encoder -------
  7. ### CNN-Backbone
  8. 'backbone': 'elannet',
  9. 'pretrained': False,
  10. 'bk_act': 'silu',
  11. 'bk_norm': 'BN',
  12. 'bk_dpw': False,
  13. 'width': 0.25,
  14. 'depth': 0.34,
  15. 'stride': [8, 16, 32], # P3, P4, P5
  16. 'max_stride': 32,
  17. ### CNN-Neck
  18. 'neck': 'sppf',
  19. 'neck_expand_ratio': 0.5,
  20. 'pooling_size': 5,
  21. 'neck_act': 'silu',
  22. 'neck_norm': 'BN',
  23. 'neck_depthwise': False,
  24. ### CNN-CSFM
  25. 'fpn': 'yolovx_pafpn',
  26. 'fpn_reduce_layer': 'conv',
  27. 'fpn_downsample_layer': 'conv',
  28. 'fpn_core_block': 'elanblock',
  29. 'fpn_act': 'silu',
  30. 'fpn_norm': 'BN',
  31. 'fpn_depthwise': False,
  32. ## ------- Memory Decoder -------
  33. 'dim_compressed': 300,
  34. 'com_dim_feedforward': 1024,
  35. 'com_num_heads': 8,
  36. 'com_dropout': 0.1,
  37. 'com_act': 'silu',
  38. ## ------- Transformer Decoder -------
  39. 'd_model': 256,
  40. 'attn_type': 'mhsa',
  41. 'num_decoder_layers': 6,
  42. 'num_queries': 300,
  43. 'de_dim_feedforward': 1024,
  44. 'de_num_heads': 8,
  45. 'de_dropout': 0.1,
  46. 'de_act': 'silu',
  47. 'de_norm': 'LN',
  48. # ---------------- Train config ----------------
  49. ## input
  50. 'multi_scale': [0.5, 1.0], # 320 -> 640
  51. 'trans_type': 'yolov5_nano',
  52. # ---------------- Assignment config ----------------
  53. ## matcher
  54. 'set_cost_class': 2.0,
  55. 'set_cost_bbox': 5.0,
  56. 'set_cost_giou': 2.0,
  57. # ---------------- Loss config ----------------
  58. ## loss weight
  59. 'focal_alpha': 0.25,
  60. 'loss_cls_weight': 1.0,
  61. 'loss_box_weight': 5.0,
  62. 'loss_giou_weight': 2.0,
  63. # ---------------- Train config ----------------
  64. 'trainer_type': 'detr',
  65. },
  66. }