🤬
Revision indexing in progress... (symbol navigation in revisions will be accurate after indexed)
  • ■ ■ ■ ■ ■
    migration/129_tasklog.up.sql
    skipped 19 lines
    20 20   `created_at` datetime DEFAULT NULL,
    21 21   `updated_at` datetime DEFAULT NULL,
    22 22   `id` varchar(255) NOT NULL,
     23 + `name` varchar(255) DEFAULT NULL,
    23 24   `task` varchar(255) DEFAULT NULL,
    24 25   `task_log_id` varchar(255) DEFAULT NULL,
    25 26   `cluster_id` varchar(255) DEFAULT NULL,
    skipped 19 lines
  • ■ ■ ■ ■ ■
    pkg/dto/cluster_import.go
    skipped 183 lines
    184 184   EtcdCompactionRetention: c.KoClusterInfo.EtcdCompactionRetention,
    185 185   EtcdMaxRequest: c.KoClusterInfo.EtcdMaxRequest,
    186 186   EtcdQuotaBackend: c.KoClusterInfo.EtcdQuotaBackend,
     187 + KubeRouter: c.Router,
    187 188   LbMode: c.KoClusterInfo.LbMode,
    188 189   LbKubeApiserverIp: c.KoClusterInfo.LbKubeApiserverIp,
    189 190   KubeApiServerPort: c.KoClusterInfo.KubeApiServerPort,
    skipped 44 lines
  • ■ ■ ■ ■ ■
    pkg/dto/tasklog.go
    skipped 5 lines
    6 6   
    7 7  type TaskLog struct {
    8 8   model.TaskLog `json:"tasklogs"`
     9 + Name string `json:"name"`
    9 10  }
    10 11   
    11 12  type Logs struct {
    skipped 3 lines
  • ■ ■ ■ ■ ■
    pkg/model/task_log.go
    skipped 23 lines
    24 24  type TaskLogDetail struct {
    25 25   common.BaseModel
    26 26   ID string `json:"id"`
     27 + Name string `json:"name"`
    27 28   Task string `json:"task"`
    28 29   TaskLogID string `json:"taskLogID"`
    29 30   ClusterID string `json:"clusterID"`
    skipped 36 lines
  • ■ ■ ■ ■ ■ ■
    pkg/service/cluster/adm/facts/default.go
    skipped 86 lines
    87 87   EnableNpdFactName = "enable_npd"
    88 88   EnableIstioFactName = "enable_istio"
    89 89   
    90  - EnableNfsFactName = "enable_nfs_provisioner"
    91  - EnableGfsFactName = "enable_gfs_provisioner"
    92  - EnableCephBlockFactName = "enable_external_ceph_block_provisioner"
    93  - EnableCephFsFactName = "enable_external_cephfs_provisioner"
    94  - EnableCinderFactName = "enable_cinder_provisioner"
    95  - EnableVsphereFactName = "enable_vsphere_provisioner"
    96  - EnableOceanstorFactName = "enable_oceanstor_provisioner"
    97  - EnableRookFactName = "enable_rook_provisioner"
     90 + ProvisionerNamespaceFactName = "provisioner_namespace"
     91 + EnableNfsFactName = "enable_nfs_provisioner"
     92 + EnableGfsFactName = "enable_gfs_provisioner"
     93 + EnableCephBlockFactName = "enable_external_ceph_block_provisioner"
     94 + EnableCephFsFactName = "enable_external_cephfs_provisioner"
     95 + EnableCinderFactName = "enable_cinder_provisioner"
     96 + EnableVsphereFactName = "enable_vsphere_provisioner"
     97 + EnableOceanstorFactName = "enable_oceanstor_provisioner"
     98 + EnableRookFactName = "enable_rook_provisioner"
    98 99  )
    99 100   
    100 101  var DefaultFacts = map[string]string{
    skipped 73 lines
    174 175   EnableDnsCacheFactName: "enable",
    175 176   EnableNpdFactName: "disable",
    176 177   EnableIstioFactName: "disable",
     178 + 
     179 + ProvisionerNamespaceFactName: "kube-system",
    177 180  }
    178 181   
  • ■ ■ ■ ■ ■ ■
    pkg/service/cluster_import.go
    skipped 34 lines
    35 35   projectRepository repository.ProjectRepository
    36 36   projectResourceRepository repository.ProjectResourceRepository
    37 37   msgService MsgService
     38 + componentService ComponentService
    38 39  }
    39 40   
    40 41  func NewClusterImportService() *clusterImportService {
    skipped 2 lines
    43 44   projectRepository: repository.NewProjectRepository(),
    44 45   projectResourceRepository: repository.NewProjectResourceRepository(),
    45 46   msgService: NewMsgService(),
     47 + componentService: NewComponentService(),
    46 48   }
    47 49  }
    48 50   
    skipped 151 lines
    200 202   return fmt.Errorf("can not save tool %s", err.Error())
    201 203   }
    202 204   }
     205 + subscribe := model.NewMsgSubscribe(constant.ClusterOperator, constant.Cluster, cluster.ID)
     206 + if err := tx.Create(&subscribe).Error; err != nil {
     207 + c.handlerImportError(tx, cluster, err)
     208 + }
    203 209   
    204 210   subscribe := model.NewMsgSubscribe(constant.ClusterOperator, constant.Cluster, cluster.ID)
    205 211   if err := tx.Create(&subscribe).Error; err != nil {
    skipped 25 lines
    231 237   hostService := NewHostService()
    232 238   go func() {
    233 239   _ = hostService.SyncList(synchosts)
     240 + }()
     241 + go func() {
     242 + syncList := []string{"metrics-server", "traefik", "ingress-nginx", "gpu", "dns-cache", "istio", "npd"}
     243 + syncDTO := &dto.ComponentSync{ClusterName: cluster.Name, Names: syncList}
     244 + _ = c.componentService.Sync(syncDTO)
    234 245   }()
    235 246   return nil
    236 247  }
    skipped 310 lines
  • ■ ■ ■ ■ ■ ■
    pkg/service/cluster_storage_provisioner.go
    skipped 125 lines
    126 126   playbook := c.loadPlayBookName(dp, creation.Vars)
    127 127   task := model.TaskLogDetail{
    128 128   ID: dp.ID,
     129 + Name: fmt.Sprintf("%s (%s)", dp.Name, dp.Type),
    129 130   Task: fmt.Sprintf("%s (%s)", playbook, constant.StatusEnabled),
    130 131   ClusterID: cluster.ID,
    131 132   LastProbeTime: time.Now().Unix(),
    skipped 67 lines
    199 200   playbook := c.loadPlayBookName(provisioner, Vars)
    200 201   task := model.TaskLogDetail{
    201 202   ID: fmt.Sprintf("%s (%s)", provisioner.ID, constant.StatusDisabled),
     203 + Name: fmt.Sprintf("%s (%s)", provisioner.Name, provisioner.Type),
    202 204   Task: fmt.Sprintf("%s (%s)", playbook, constant.StatusDisabled),
    203 205   ClusterID: cluster.ID,
    204 206   LastProbeTime: time.Now().Unix(),
    skipped 88 lines
    293 295   }
    294 296   c.changeStatus(provisioner, constant.StatusRunning, nil)
    295 297   case "rook-ceph":
    296  - if err := phases.WaitForDeployRunning(provisioner.Namespace, "rook-ceph-operator", client); err != nil {
     298 + if err := phases.WaitForDeployRunning("rook-ceph", "rook-ceph-operator", client); err != nil {
    297 299   c.changeStatus(provisioner, constant.StatusFailed, err)
    298 300   continue
    299 301   }
    skipped 91 lines
    391 393   if err != nil {
    392 394   return admCluster, writer, err
    393 395   }
     396 + admCluster.Kobe.SetVar(facts.ProvisionerNamespaceFactName, provisioner.Namespace)
    394 397   
    395 398   switch provisioner.Type {
    396 399   case "nfs":
    397 400   admCluster.Kobe.SetVar(facts.EnableNfsFactName, operation)
    398  - case "gfs":
     401 + case "glusterfs":
    399 402   admCluster.Kobe.SetVar(facts.EnableGfsFactName, operation)
    400 403   case "external-ceph-block":
    401  - fmt.Println(facts.EnableCephBlockFactName, operation)
    402 404   admCluster.Kobe.SetVar(facts.EnableCephBlockFactName, operation)
    403 405   case "external-cephfs":
    404 406   admCluster.Kobe.SetVar(facts.EnableCephFsFactName, operation)
    skipped 12 lines
  • ■ ■ ■ ■ ■ ■
    pkg/service/component.go
    skipped 160 lines
    161 161   playbook := c.loadPlayBookName(creation.Name)
    162 162   task := model.TaskLogDetail{
    163 163   ID: component.ID,
     164 + Name: creation.Name,
    164 165   Task: fmt.Sprintf("%s (%s)", playbook, constant.StatusEnabled),
    165 166   ClusterID: cluster.ID,
    166 167   LastProbeTime: time.Now().Unix(),
    skipped 52 lines
    219 220   playbook := c.loadPlayBookName(name)
    220 221   task := model.TaskLogDetail{
    221 222   ID: fmt.Sprintf("%s (%s)", component.ID, constant.StatusDisabled),
     223 + Name: component.Name,
    222 224   Task: fmt.Sprintf("%s (%s)", playbook, constant.StatusDisabled),
    223 225   ClusterID: cluster.ID,
    224 226   LastProbeTime: time.Now().Unix(),
    skipped 238 lines
  • ■ ■ ■ ■ ■
    pkg/service/task_log.go
    skipped 98 lines
    99 99   StartTime: tasklogs[t].StartTime,
    100 100   EndTime: tasklogs[t].EndTime,
    101 101   },
     102 + Name: tasklogs[t].Name,
    102 103   })
    103 104   }
    104 105   p.Items = datas
    skipped 273 lines
Please wait...
Page is in error, reload to recover