Hello Davide! I really appreciate your kind help!
kubectl get networkpolicy gave the following: "No resources found in default namespace." Actually I've tried to run the Kubernetes plugin in a so-called "some" namespace, but there is no networkpolicy in "some" NS neither. (I've changed the name of the NS to some.) kubectl get networkpolicy -n some No resources found in some namespace. Best regards, Zsolt On Thu, Jan 12, 2023 at 8:00 AM Davide F. <bacula-...@dflc.ch> wrote: > Hello Zsolt, > > Indeed, the NetworkPolicy you've provided doesn't seem to be involved in > the issue your facing. > > Let's keep trying to figure out what's going on with your setup > > Could you run these commands below > > kubectl get networkpolicy > > > and if you get some result, run > > kubectl describe networkpolicy <networkpolicy-name> > > > In the meantime, I'll setup a "test" environement and see if I'm facing > the problem. > > I'll keep you updated. > > Best regards > > Davide > > On Wed, Jan 11, 2023 at 5:54 PM Zsolt Kozak <koza...@gmail.com> wrote: > >> Hi! >> >> Yes, but only one tiny: >> >> kind: NetworkPolicy >> apiVersion: networking.k8s.io/v1 >> metadata: >> name: allow-apiserver >> namespace: calico-apiserver >> ownerReferences: >> - apiVersion: operator.tigera.io/v1 >> kind: APIServer >> name: default >> controller: true >> blockOwnerDeletion: true >> managedFields: >> - manager: operator >> operation: Update >> apiVersion: networking.k8s.io/v1 >> spec: >> podSelector: >> matchLabels: >> apiserver: 'true' >> ingress: >> - ports: >> - protocol: TCP >> port: 5443 >> policyTypes: >> - Ingress >> status: {} >> >> But I guess it's an allow, not a block policy. (I'm somewhat new to >> Kubernetes and not too familiar with network policies...) >> >> Best regards, >> Zsolt >> >> On Wed, Jan 11, 2023 at 5:47 PM Davide F. <bacula-...@dflc.ch> wrote: >> >>> Hi, >>> >>> Are you using some kind of network policy within your cluster ? >>> >>> Best, >>> >>> Davide >>> >>> On Wed, 11 Jan 2023 at 10:53 Zsolt Kozak <koza...@gmail.com> wrote: >>> >>>> Hello Davide! >>>> >>>> I am running the File Daemon on the master node, on the control plane. >>>> It's Kubernetes vanilla, version 1.25.4. >>>> No, the master node is running on the same subnet as the workers. >>>> >>>> It should be some network issue, I think. >>>> >>>> Best regards, >>>> Zsolt >>>> >>>> On Wed, Jan 11, 2023 at 8:45 AM Davide F. <bacula-...@dflc.ch> wrote: >>>> >>>>> Hello Kozak, >>>>> >>>>> I haven’t tried k8s plugin but let me try to understand what could be >>>>> the root cause of your problem. >>>>> >>>>> Could you explain further point 1 please ? >>>>> On which node are you running the file daemon ? >>>>> >>>>> Which version / flavor of Kubernetes are you using ? >>>>> >>>>> Is it Kubernetes vanilla ? OpenShift ? Tansu ? >>>>> >>>>> Depending on your feedback from the first question, does master nodes >>>>> runs in a different subnet than worker’s ? >>>>> >>>>> Thanks for your feedback >>>>> >>>>> Best, >>>>> >>>>> Davide >>>>> >>>>> On Tue, 10 Jan 2023 at 21:12 Zsolt Kozak <koza...@gmail.com> wrote: >>>>> >>>>>> Hello, >>>>>> >>>>>> I have some problems with backuping Kubernetes PVCs with Bacula >>>>>> Kubernetes Plugin. (I have asked it on bacula-users mailing list but got >>>>>> no >>>>>> answer.) >>>>>> >>>>>> I am using the latest 13.0.1 Bacula from the community builds on >>>>>> Debian Bullseye hosts. >>>>>> >>>>>> Backuping only the Kubernetes objects except Persistent Volume Claims >>>>>> (PVC) works like a charm. I've installed the Kubernetes plugin and the >>>>>> latest Bacula File Daemon on the master node (control plane) of our >>>>>> Kubernetes cluster. Bacula can access the Kubernetes cluster and backup >>>>>> every single object as YAML files. >>>>>> >>>>>> The interesting part comes with trying to backup a PVC... >>>>>> >>>>>> First of all I could build my own Bacula Backup Proxy Pod Image from >>>>>> the source and it's deployed into our local Docker image repository >>>>>> (repo). >>>>>> The Bacula File Daemon is configured properly I guess. Backup process >>>>>> started and the following things happened. >>>>>> >>>>>> 1. Bacula File Daemon deployed Bacula Backup Proxy Pod Image into the >>>>>> Kubernetes cluster, so Bacula-backup container pod started. >>>>>> 2. I got into the pod and I could see the Baculatar application >>>>>> started and running. >>>>>> 3. The k8s_backend application started on the Bacula File Daemon host >>>>>> (kubernetes.server) in 2 instances. >>>>>> 4. From the Bacula-backup pod I could check that Baculatar could >>>>>> connect to the k8s_backend at the default 9104 port >>>>>> (kubernetes.server:9104). >>>>>> 5. I checked the console messages of the job with Bat that Bacula >>>>>> File Daemon started to process the configured PVC, started to write a >>>>>> pvc.tar but nothing happened. >>>>>> 6. After default 600 sec, after timeout the job was cancelled. >>>>>> 7. It may be important that Bacula File Daemon could not delete the >>>>>> Bacula-backup pod. (It could create it but could not delete it.) >>>>>> >>>>>> >>>>>> Could you please tell me what's wrong? >>>>>> >>>>>> >>>>>> Here are some log parts. (I've changed some sensitive data.) >>>>>> >>>>>> >>>>>> Bacula File Daemon configuration: >>>>>> >>>>>> FileSet { >>>>>> Name = "Kubernetes Set" >>>>>> Include { >>>>>> Options { >>>>>> signature = SHA512 >>>>>> compression = GZIP >>>>>> Verify = pins3 >>>>>> } >>>>>> Plugin = "kubernetes: \ >>>>>> debug=1 \ >>>>>> baculaimage=repo/bacula-backup:04jan23 \ >>>>>> namespace=namespace \ >>>>>> pvcdata \ >>>>>> pluginhost=kubernetes.server \ >>>>>> timeout=120 \ >>>>>> verify_ssl=0 \ >>>>>> fdcertfile=/etc/bacula/certs/bacula-backup.cert \ >>>>>> fdkeyfile=/etc/bacula/certs/bacula-backup.key" >>>>>> } >>>>>> } >>>>>> >>>>>> >>>>>> >>>>>> Bacula File Daemon debug log (parts): >>>>>> >>>>>> >>>>>> DEBUG:[baculak8s/jobs/estimation_job.py:134 in processing_loop] >>>>>> processing get_annotated_namespaced_pods_data:namespace:nrfound:0 >>>>>> DEBUG:[baculak8s/plugins/kubernetes_plugin.py:319 in >>>>>> list_pvcdata_for_namespace] list pvcdata for namespace:namespace >>>>>> pvcfilter=True estimate=False >>>>>> DEBUG:[baculak8s/plugins/k8sbackend/pvcdata.py:108 in >>>>>> pvcdata_list_namespaced] pvcfilter: True >>>>>> DEBUG:[baculak8s/plugins/k8sbackend/pvcdata.py:112 in >>>>>> pvcdata_list_namespaced] found:some-claim >>>>>> DEBUG:[baculak8s/plugins/k8sbackend/pvcdata.py:127 in >>>>>> pvcdata_list_namespaced] add pvc: {'name': 'some-claim', 'node_name': >>>>>> None, >>>>>> 'storage_class_name': 'nfs-client', 'capacity': '2Gi', 'fi': >>>>>> <baculak8s.entities.file_info.FileInfo object at 0x7ffaa55bfcc0>} >>>>>> DEBUG:[baculak8s/jobs/estimation_job.py:165 in processing_loop] >>>>>> processing list_pvcdata_for_namespace:namespace:nrfound:1 >>>>>> DEBUG:[baculak8s/jobs/estimation_job.py:172 in processing_loop] >>>>>> PVCDATA:some-claim:{'name': 'some-claim', 'node_name': 'node1', >>>>>> 'storage_class_name': 'nfs-client', 'capacity': '2Gi', 'fi': >>>>>> <baculak8s.entities.file_info.FileInfo object at 0x7ffaa55bfcc0>} >>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet >>>>>> I000041 >>>>>> Start backup volume claim: some-claim >>>>>> >>>>>> DEBUG:[baculak8s/jobs/job_pod_bacula.py:298 in prepare_bacula_pod] >>>>>> prepare_bacula_pod:token=xx88M5oggQJ....4YDbSwBRxTOhT namespace=namespace >>>>>> DEBUG:[baculak8s/jobs/job_pod_bacula.py:136 in prepare_pod_yaml] >>>>>> pvcdata: {'name': 'some-claim', 'node_name': 'node1', >>>>>> 'storage_class_name': >>>>>> 'nfs-client', 'capacity': '2Gi', 'fi': >>>>>> <baculak8s.entities.file_info.FileInfo object at 0x7ffaa55bfcc0>} >>>>>> DEBUG:[baculak8s/plugins/k8sbackend/baculabackup.py:102 in >>>>>> prepare_backup_pod_yaml] host:kubernetes.server port:9104 >>>>>> namespace:namespace image:repo/bacula-backup:04jan23 >>>>>> job:KubernetesBackup.2023-01-04_21.05.03_10:410706 >>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet >>>>>> I000149 >>>>>> Prepare Bacula Pod on: node1 with: repo/bacula-backup:04jan23 >>>>>> <IfNotPresent> kubernetes.server:9104 >>>>>> >>>>>> DEBUG:[baculak8s/jobs/job_pod_bacula.py:198 in >>>>>> prepare_connection_server] prepare_connection_server:New >>>>>> ConnectionServer: >>>>>> 0.0.0.0:9104 >>>>>> DEBUG:[baculak8s/util/sslserver.py:180 in listen] >>>>>> ConnectionServer:Listening... >>>>>> DEBUG:[baculak8s/jobs/job_pod_bacula.py:307 in prepare_bacula_pod] >>>>>> prepare_bacula_pod:start pod >>>>>> INFO:[baculak8s/plugins/kubernetes_plugin.py:771 in >>>>>> backup_pod_isready] backup_pod_status:isReady: False / 0 >>>>>> INFO:[baculak8s/plugins/kubernetes_plugin.py:771 in >>>>>> backup_pod_isready] backup_pod_status:isReady: True / 1 >>>>>> DEBUG:[baculak8s/jobs/estimation_job.py:183 in _estimate_file] >>>>>> {'name': 'some-claim', 'node_name': 'node1', 'storage_class_name': >>>>>> 'nfs-client', 'capacity': '2Gi', 'fi': >>>>>> <baculak8s.entities.file_info.FileInfo object at 0x7ffaa55bfcc0>} >>>>>> DEBUG:[baculak8s/jobs/estimation_job.py:190 in _estimate_file] >>>>>> file_info: {FileInfo >>>>>> name:/@kubernetes/namespaces/namespace/persistentvolumeclaims/some-claim.tar >>>>>> namespace:None type:F objtype:pvcdata cached:False} >>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet >>>>>> C000079 >>>>>> >>>>>> FNAME:/@kubernetes/namespaces/namespace/persistentvolumeclaims/some-claim.tar >>>>>> >>>>>> >>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet >>>>>> C000040 >>>>>> TSTAMP:1672861077 1672861077 1672861077 >>>>>> >>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet >>>>>> C000031 >>>>>> STAT:F 2147483648 0 0 100640 1 >>>>>> >>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet >>>>>> F000000 >>>>>> (EOD PACKET) >>>>>> >>>>>> DEBUG:[baculak8s/jobs/backup_job.py:77 in __backup_pvcdata] >>>>>> backup_pvcdata:data recv >>>>>> DEBUG:[baculak8s/io/log.py:110 in save_sent_packet] Sent Packet >>>>>> C000005 >>>>>> DATA >>>>>> >>>>>> DEBUG:[baculak8s/util/sslserver.py:193 in handle_connection] >>>>>> ConnectionServer:Connection from: ('192.168.XX.YY', 10541) >>>>>> DEBUG:[baculak8s/util/sslserver.py:145 in gethello] ['Hello', >>>>>> 'KubernetesBackup.2023-01-04_21.05.03_10', '410706'] >>>>>> DEBUG:[baculak8s/util/token.py:57 in check_auth_data] >>>>>> AUTH_DATA:Token: xx88M5oggQJuGsPbtD........ohQjeU7PkA4YDbSwBRxTOhT >>>>>> DEBUG:[baculak8s/util/token.py:59 in check_auth_data] >>>>>> RECV_TOKEN_DATA:Token: xx88M5oggQJuGsPbtD....ohQjeU7PkA4YDbSwBRxTOhT >>>>>> DEBUG:[baculak8s/util/sslserver.py:105 in authenticate] >>>>>> ConnectionServer:Authenticated >>>>>> >>>>>> .... after timeout >>>>>> >>>>>> DEBUG:[baculak8s/jobs/job_pod_bacula.py:121 in handle_pod_data_recv] >>>>>> handle_pod_data_recv:EOT >>>>>> DEBUG:[baculak8s/util/sslserver.py:201 in handle_connection] >>>>>> ConnectionServer:Finish - disconnect. >>>>>> DEBUG:[baculak8s/jobs/backup_job.py:85 in __backup_pvcdata] >>>>>> backup_pvcdata:logs recv >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Job messages: >>>>>> >>>>>> bacula-dir No prior or suitable Full backup found in catalog for the >>>>>> current FileSet. Doing FULL backup. >>>>>> The FileSet "Kubernetes Set" was modified on 2023-01-04 20:20:41, >>>>>> this is after the last successful backup on 2023-01-04 19:19:49. >>>>>> bacula-sd Ready to append to end of Volume "Full-XXX" >>>>>> size=3,838,161,002 >>>>>> bacula-fd Connected to Storage at bacula.server:9103 with TLS >>>>>> bacula-sd Volume "Full-XXXX" previously written, moving to end of >>>>>> data. >>>>>> bacula-dir Connected to Client "bacula-fd" at kubernetes.server:9102 >>>>>> with TLS >>>>>> Using Device "FileStorageEeyoreFull" to write. >>>>>> Connected to Storage "InternalStorageFull" at bacula.server:9103 >>>>>> with TLS >>>>>> Start Backup JobId 410830, >>>>>> Job=KubernetesBackup.2023-01-04_21.05.03_10 >>>>>> bacula-fd kubernetes: Prepare Bacula Pod on: node with: >>>>>> repo/bacula-backup:04jan23 kubernetes.server:9104 >>>>>> kubernetes: Processing namespace: namespace >>>>>> kubernetes: Start backup volume claim: some-claim >>>>>> kubernetes: Connected to Kubernetes 1.25 - v1.25.4. >>>>>> bacula-dir >>>>>> Error: Bacula Enterprise bacula-dir 13.0.1 (05Aug22): >>>>>> Build OS: x86_64-pc-linux-gnu-bacula-enterprise >>>>>> debian 11.2 >>>>>> JobId: 410830 >>>>>> Job: KubernetesBackup.2023-01-04_21.05.03_10 >>>>>> Backup Level: Full (upgraded from Differential) >>>>>> Client: "bacula-fd" 13.0.1 (05Aug22) >>>>>> x86_64-pc-linux-gnu-bacula-enterprise,debian,10.11 >>>>>> FileSet: "Kubernetes Set" 2023-01-04 20:20:41 >>>>>> Pool: "Full-Pool" (From Job FullPool override) >>>>>> Catalog: "MyCatalog" (From Client resource) >>>>>> Storage: "InternalStorageFull" (From Pool resource) >>>>>> Scheduled time: 04-Jan-2023 21:05:03 >>>>>> Start time: 04-Jan-2023 21:27:04 >>>>>> End time: 04-Jan-2023 21:29:06 >>>>>> Elapsed time: 2 mins 2 secs >>>>>> Priority: 10 >>>>>> FD Files Written: 23 >>>>>> SD Files Written: 0 >>>>>> FD Bytes Written: 52,784 (52.78 KB) >>>>>> SD Bytes Written: 0 (0 B) >>>>>> Rate: 0.4 KB/s >>>>>> Software Compression: 100.0% 1.0:1 >>>>>> Comm Line Compression: 5.6% 1.1:1 >>>>>> Snapshot/VSS: no >>>>>> Encryption: yes >>>>>> Accurate: yes >>>>>> Volume name(s): Full-XXXX >>>>>> Volume Session Id: 43 >>>>>> Volume Session Time: 1672853724 >>>>>> Last Volume Bytes: 3,838,244,105 (3.838 GB) >>>>>> Non-fatal FD errors: 3 >>>>>> SD Errors: 0 >>>>>> FD termination status: OK >>>>>> SD termination status: SD despooling Attributes >>>>>> Termination: *** Backup Error *** >>>>>> Fatal error: catreq.c:680 Restore object create error. >>>>>> bacula-fd >>>>>> Error: kubernetes: PTCOMM cannot get packet header from backend. >>>>>> bacula-dir Fatal error: sql_create.c:1273 Create db Object record >>>>>> INSERT INTO RestoreObject >>>>>> (ObjectName,PluginName,RestoreObject,ObjectLength,ObjectFullLength,ObjectIndex,ObjectType,ObjectCompression,FileIndex,JobId) >>>>>> VALUES ('RestoreOptions','kubernetes: \n debug=1 \n >>>>>> baculaimage=repo/bacula-backup:04jan23 \n >>>>>> namespace=namespace \n pvcdata \n >>>>>> pluginhost=kubernetes.server \n timeout=120 \n >>>>>> verify_ssl=0 \n >>>>>> fdcertfile=/etc/bacula/certs/bacula-backup.cert >>>>>> \n >>>>>> fdkeyfile=/etc/bacula/certs/bacula-backup.key','# Plugin >>>>>> configuration file\n# Version 1\nOptPrompt=\"K8S config >>>>>> file\"\nOptDefault=\"*None*\"\nconfig=@STR@\n\n >>>>>> OptPrompt=\"K8S API server >>>>>> URL/Host\"\nOptDefault=\"*None*\"\nhost=@STR@\n\nOptPrompt=\"K8S >>>>>> Bearertoken\"\nOptDefault=\"*None*\"\ntoken=@STR@\n\nOptPrompt=\"K8S >>>>>> API server cert verification\"\n >>>>>> OptDefault=\"True\"\nverify_ssl=@BOOL@\n\nOptPrompt=\"Custom CA >>>>>> Certs file to >>>>>> use\"\nOptDefault=\"*None*\"\nssl_ca_cert=@STR@\n\nOptPrompt=\"Output >>>>>> format when saving to file (JSON, YAML)\"\n >>>>>> OptDefault=\"RAW\"\noutputformat=@STR@\n\nOptPrompt=\"The address >>>>>> for listen to incoming backup pod >>>>>> data\"\nOptDefault=\"*FDAddress*\"\nfdaddress=@STR@\n\n >>>>>> OptPrompt=\"The port for opening socket for >>>>>> listen\"\nOptDefault=\"9104\"\nfdport=@INT32@\n\nOptPrompt=\"The >>>>>> endpoint address for backup pod to connect\"\n >>>>>> OptDefault=\"*FDAddress*\"\npluginhost=@STR@\n\nOptPrompt=\"The >>>>>> endpoint port to connect\"\nOptDefault=\"9104\"\n >>>>>> pluginport=@INT32@\n\n',859,859,0,27,0,1,410830) failed. ERR=Data >>>>>> too long for column 'PluginName' at row 1 >>>>>> >>>>>> bacula-sd Sending spooled attrs to the Director. Despooling 8,214 >>>>>> bytes ... >>>>>> bacula-fd >>>>>> Error: kubernetes: Error closing backend. Err=Child exited with code 1 >>>>>> Fatal error: kubernetes: Wrong backend response to JobEnd command. >>>>>> bacula-sd Elapsed time=00:02:02, Transfer rate=659 Bytes/second >>>>>> bacula-fd >>>>>> Error: kubernetes: PTCOMM cannot get packet header from backend. >>>>>> >>>>>> Error: kubernetes: Cannot successfully start bacula-backup pod in >>>>>> expected time! >>>>>> >>>>>> Error: kubernetes: Job already running in 'namespace' namespace. >>>>>> Check logs or delete bacula-backup Pod manually. >>>>>> >>>>>> >>>>>> Best regards, >>>>>> Zsolt >>>>>> >>>>>> _______________________________________________ >>>>>> Bacula-devel mailing list >>>>>> Bacula-devel@lists.sourceforge.net >>>>>> https://lists.sourceforge.net/lists/listinfo/bacula-devel >>>>>> >>>>>
_______________________________________________ Bacula-devel mailing list Bacula-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/bacula-devel