Tag Archives: Operation and maintenance

Linux: How to Solve pyperclip Module Error

When pyperclip is used, the error message is as follows:

xclip needs to be installed
Step 1: install the epel repository

# yum install epel-release.noarch
             or
# dnf install epel-release

Step 2: install xclip from the EPEL repository

# dnf install xclip
Last metadata expiration check: 0:06:09 ago on Mon 07 Feb 2022 10:39:39 AM IST.
Dependencies resolved.
=============================================================================================================================================================
 Package                             Architecture                         Version                                   Repository                          Size
=============================================================================================================================================================
Installing:
 xclip                               x86_64                               0.13-8.el8                                epel                                35 k

Transaction Summary
=============================================================================================================================================================
Install  1 Package

After installation, you can use the pyperclip module normally.

How to Solve Qualcomm QFIL9008 port brush error

 

ERROR: function: sahara_rx_data:276 Unable to read packet header. Only read 0 bytes.


ERROR: function: sahara_main:982 Sahara protocol error


ERROR: function: main:320 Uploading  Image using Sahara protocol failed

I use the latest version QPST.2.7.496. After downloading, the above problems have occurred all the time. Later, I changed to the old version, it is OK.

In addition, you can turn off the firewall

Or disable driver signature: restart, press F8 repeatedly during startup, and select disable driver signature enforcement

[Solved] openssl error: libcrypto.so.1.0.0: cannot open shared object file: Error 40

Problem Description: start to execute OpenSSL and report an error:

openssl

openssl: error while loading shared libraries: libcrypto.so.1.0.0: cannot open shared object file: No such file or directory

Then find the corresponding file through the search instruction, and soft connect it

find/-name libcrypto.so.1.0.0    find files
/opt/acb/lib/libcrypto.so.1.0.0   search the outcome
ln -s libcrypto.so.1.0.0 /usr/lib64/  soft connect

However, it is found that the implementation of OpenSSL still reports errors, as follows

openssl
openssl: error while loading shared libraries: libcrypto.so.1.0.0: cannot open shared object file: Error 40

Check libcrypto.so.1.0.0 that reported error, and find that libcrypto.so.1.0.0 is also a soft connection

lrwxrwxrwx 1 splunk root      15 Jan  9  2020 libssl.so -> libssl.so.1.0.0
-r-xr-xr-x 2 splunk root  485304 Jan  9  2020 libssl.so.1.0.0

The solution is to connect the source file directly and solve the problem

ln -s libcrypto.so /usr/lib64/libcrypto.so.1.0.0

[Solved] error while loading shared libraries: libjson.so.0: cannot open shared object file: No such file or

error while loading shared libraries: libjson.so.0: cannot open shared object file: No such file or

Solution:

sudo vim  /etc/ld.so.conf

Add /usr/local/lib to the file, which is the path is libjson.so.0 located

include /etc/ld.so.conf.d/*.conf 
/usr/local/lib

Finally, the terminal inputs sudo ldconfig

[Solved] Readiness probe failed: calico/node is not ready: BIRD is not ready: Error querying BIRD: unable to

Cluster calico deployed by kuberadm runs incorrectly

report errors:

1, check calico

kubectl get pods -n kube-system

Calico-node 0/1 Running

2. check the log

kubectl describe calico-node-gdv9r -n kube-system

Calico error:
Readiness probe failed: calico/node is not ready: BIRD is not ready: Error querying BIRD: unable to connect to BIRDv4 socket: dial unix /var/run/bird/bird.ctl: connect: no such file or directory

Solution:

1. Remove the DOWN NIC/interface that Docker and Calico errors.

ip link docker0

2, delete Calico pod (pod will be rebuilt and run by k8s cluster after deletion)

kubectl delete pod calico-node-gdv9r -n kube-system

3, check Calico again and find it is normal</ font>

kubectl get pods -n kube-system

[Solved] Xshelln Connect to Local VM linux Error: Could not connect to ‘XXX.168.122.1‘ (port 22): Connection failed

The error is shown in the following picture:

Solution:

Method 1:

Restart VMware, xsehell!

The second method:

1. Start SSH

service sshd start

2. Set SSH to start automatically

chkconfigsshd on

Note: if there is no SSH, install SSH first, and then perform the previous steps 1 and 2

yum install openssh-server

Finally, we go to xshell to test the connection

Failed to remove multipath map 320b508ca45022b80 [How to Solve]

Failed to remove multipath map 320b508ca45022b80

1. Project scenario

Host os:kylin-server-10-sp1-release-build02-20210518-arm64
docker:docker-ce-18.09.7
cloud: openstack queens
storage: same acs5000
VM os: kylin-server-10-sp1-release-build02-20210518-arm64


2. Problem description and cause analysis

2.1 problem description

The volume-based virtual machine can be created normally, but the error after restarting the virtual machine, checking the logs of nova-compute, found that it reports ProcessExecutionError:unexpected error while running command. command: multipah -f 320b508ca45022b80 failed, map in use, failed to remove multipath map 320b508ca45022b80.
I manually executed multipah -f 320b508ca45022b80, and it did report the status of in use, so I suspect that there are processes using the volume, and I found that the same volume group name was activated through lvdisplay, vgdisplay and lsblk, so I suspect that the virtual machine and the physical machine used the same volume group name, and the volume group name was activated after the virtual machine started. The VM has been activated, and the process of reactivating all logical volumes in the volume group failed, resulting in multipath -f failure. Therefore, we need to configure lvm to activate only the logical volumes of the system, check the system volumes by lsblk, and then configure accordingly, edit /etc/lvm/lvm.conf and modify the following content

devices {
        filter = [ "a/sda/", "r/.*/" ]
}
allocation {
       volume_list = ["klas"]
       auto_activation_volume_list = ["klas"]
}

Restart service:

systemctl restart lvm2-lvmetad.service lvm2-lvmetad.socket

Re create the virtual machine and restart it. It is also recommended that the virtual machine adopt other volume group names

2.2 storage configuration

2.2.1 drive

Use the same driver version zeus-driver-3.1.2.000106, copy the driver to the cinder_volume container /usr/lib/python2.7/site-packages/cinder/volume/drivers/ directory and the cinder_backup container /usr/lib/python2.7/site-packages/cinder/backup/drivers/ directory, and restart the related services.

2.2.2 configure cinder volume

vim /etc/kolla/cinder-volume/cinder.conf

[DEFAULT]
enabled_backends=toyou_ssd
[toyou_ssd]
volume_driver = cinder.volume.drivers.zeus.Acs5000_iscsi.Acs5000ISCSIDriver
san_ip = x.x.x.x
use_mutipath_for_image_xfer = True
image_volume_cache_enabled = True
san_login = cliuser
san_password = ******
acs5000_volpool_name = toyou_ssd
acs5000_target = 0
volume_backend_name = toyou_ssd

Restart the cinder-volume service. For others, please refer to the “reference scheme”


3. Solutions

View the adopted system disk through lsblk, and then edit /etc/lvm/lvm.conf to modify the following contents

devices {
        filter = [ "a/sda/", "r/.*/" ]
}
allocation {
       volume_list = ["klas"]
       auto_activation_volume_list = ["klas"]
}

Restart service:

systemctl restart lvm2-lvmetad.service lvm2-lvmetad.socket

Note that it is mainly the filter. The drive letter in the filter is determined by the system disk recognized by lsblk, which may be SDB or nvme, etc

[Solved] onlyoffice Error: error self signed certificate and download failed

When Installing nextcloud+onlyoffice, onlyoffice failed to start and report an error:


enter the container to see the error information of out.log

[root@nextcloud ~]# docker ps -a
Emulate Docker CLI using podman. Create /etc/containers/nodocker to quiet msg.
CONTAINER ID  IMAGE                                       COMMAND     CREATED       STATUS           PORTS                                        NAMES
a7c97fb93556  docker.io/onlyoffice/documentserver:latest              30 hours ago  Up 30 hours ago  0.0.0.0:8080->80/tcp, 0.0.0.0:9000->443/tcp  onlyoffice
[root@nextcloud ~]# docker exec -it a7c97fb93556 /bin/bash
Emulate Docker CLI using podman. Create /etc/containers/nodocker to quiet msg.
root@a7c97fb93556:/# cd /var/log/onlyoffice/documentserver/converter/
root@a7c97fb93556:/var/log/onlyoffice/documentserver/converter# ls
err.log  out.log-20220729
root@a7c97fb93556:/var/log/onlyoffice/documentserver/converter#

Disabling Document Server Access Authentication
Next, disable access authentication for Document Server, which by default rejects unauthenticated requests (i.e., self-signed HTTPS requests).

I am now running Document Server with Docker, using the docker exec command to log into the container.

There seems to be only the nano editor in the container, but that’s enough.

Open /etc/onlyoffice/documentserver/default.json, go down and find the rejectUnauthorized field and change its value to false.

Restart the container.
Modify default.json

root@a7c97fb93556:/var/log/onlyoffice/documentserver/converter# cd /etc/onlyoffice/
root@a7c97fb93556:/etc/onlyoffice# ls
documentserver  documentserver-example
root@a7c97fb93556:/etc/onlyoffice# cd documentserver
root@a7c97fb93556:/etc/onlyoffice/documentserver# ls
default.json              local.json  production-linux.json
development-linux.json    log4js      production-windows.json
development-mac.json      logrotate   supervisor
development-windows.json  nginx
root@a7c97fb93556:/etc/onlyoffice/documentserver# pwd
/etc/onlyoffice/documentserver
root@a7c97fb93556:/etc/onlyoffice/documentserver#

Modify as follows: “rejectunauthorized”: false

                     "requestDefaults": {
                                "headers": {
                                        "User-Agent": "Node.js/6.13",
                                        "Connection": "Keep-Alive"
                                },
                                "gzip": true,
                                "rejectUnauthorized": false
                        },

Restart container

root@a7c97fb93556:/etc/onlyoffice/documentserver# exit
exit
[root@nextcloud ~]# docker stop a7c97fb93556
Emulate Docker CLI using podman. Create /etc/containers/nodocker to quiet msg.
Error: given PIDs did not die within timeout
[root@nextcloud ~]# docker start a7c97fb93556
Emulate Docker CLI using podman. Create /etc/containers/nodocker to quiet msg.
Error: unable to start container "a7c97fb93556650c83dd763f9578705a82f34b2673f9759e8d0ce62afc63e77c": container a7c97fb93556650c83dd763f9578705a82f34b2673f9759e8d0ce62afc63e77c must be in Created or Stopped state to be started: container state improper
[root@nextcloud ~]# reboot

Restart nextcloud

login as: root
[email protected]'s password:
Activate the web console with: systemctl enable --now cockpit.socket

Last login: Fri Jul 29 15:59:59 2022 from 192.168.182.1
[root@nextcloud ~]# setenforce 0
[root@nextcloud ~]# systemctl start https
Failed to start https.service: Unit https.service not found.
[root@nextcloud ~]# systemctl start httpd
Enter TLS private key passphrase for localhost:443 (RSA) : ******
[root@nextcloud ~]# docker ps -a
Emulate Docker CLI using podman. Create /etc/containers/nodocker to quiet msg.
CONTAINER ID  IMAGE                                       COMMAND     CREATED       STATUS      PORTS                                        NAMES
a7c97fb93556  docker.io/onlyoffice/documentserver:latest              31 hours ago  Created     0.0.0.0:8080->80/tcp, 0.0.0.0:9000->443/tcp  onlyoffice
[root@nextcloud ~]# docker start a7c97fb93556
Emulate Docker CLI using podman. Create /etc/containers/nodocker to quiet msg.
a7c97fb93556
[root@nextcloud ~]#

Start onlyoffice

run as prompted

[root@nextcloud ~]# sudo docker exec a7c97fb93556 sudo supervisorctl start ds:example
Emulate Docker CLI using podman. Create /etc/containers/nodocker to quiet msg.
sudo: unable to send audit message: Operation not permitted
ds:example: started

Successfully opened word document

[Solved] Emacs27.1 cscope Error: process-kill-without-query

1. Background:

Using emacs27.1 and cscope on XUbuntu22.04, I get an error when looking for a function.

process-kill-without-query
process-kill-without-query is a compiled Lisp function in `subr.el'.

(process-kill-without-query PROCESS &optional FLAG)

This function is obsolete since 22.1;
use `process-query-on-exit-flag' or `set-process-query-on-exit-flag'.

Say no query needed if PROCESS is running when Emacs is exited.
Optional second argument if non-nil says to require a query.
Value is t if a query was formerly required. 

2. Solutions

Solution idea.
Replace all containing (process-kill-without-query xxx) with
as follows.

(process-query-on-exit-flag xxx)
or
(set-process-query-on-exit-flag xxx nil)

1. Replace all the files containing process-kill-without-query in /usr/share/dictionaries-common/site-elisp directory
debian-ispell.el flyspell.el ispell.el

2. replace all process-kill-without-query under ~/.emacs.d

3. delete all .elc files with process-kill-without-query

 

[Solved] make menuconfig Error: recipe for target ‘menuconfig‘ failed

when using make menuconfig to configure Linux corl error: target ‘menuconfig’ failed to recipe.


Solution:

1. Check whether the libncurses5-dev library is installed

If it’s not installed, install with the command: sudo apt-get install libncurses5-dev

2. If it’s installed, it’s because the terminal window is too small. It’s solved by enlarging the window

 

[Solved] Docker Start nginx error: driver failed programming external connectivity on endpoint nginx

@[TOC](Docker start nginx error: Error response from daemon: driver failed programming external connectivity on endpoint nginx (6307a83f775bf316440e9b843841d808a84dded1b805b204484e41a73835f24a): (iptables failed: iptables –wait -t nat -A DOCKER -p tcp -d 0/0 –dport 80 -j DNAT –to-destination 172.17.0.6:80 ! -i docker0: iptables: No chain/target/match by that name.)

Problem Background
When starting nginx using docker, an error is reported

[root@localhost conf.d]# docker start nginx 
Error response from daemon: driver failed programming external connectivity on endpoint nginx (6307a83f775bf316440e9b843841d808a84dded1b805b204484e41a73835f24a):  (iptables failed: iptables --wait -t nat -A DOCKER -p tcp -d 0/0 --dport 80 -j DNAT --to-destination 172.17.0.6:80 ! -i docker0: iptables: No chain/target/match by that name.

Solution:

1. Before starting the nginx of docker, the firewall was closed, which led to an error in the link. You can restart docker to re-establish the link

systemctl restart docker

Restart nginx

docker start nginx