question_id
int64
82.3k
79.7M
title_clean
stringlengths
15
158
body_clean
stringlengths
62
28.5k
full_text
stringlengths
95
28.5k
tags
stringlengths
4
80
score
int64
0
1.15k
view_count
int64
22
1.62M
answer_count
int64
0
30
link
stringlengths
58
125
75,603,020
Create podman volume with custom path
I am trying to create a volume with custom path, I want the volume data path to be /home/$USER/data instead of default path which is ~/.local/share/containers/storage/volumes . I have googled it but could not find the result, so I got help from chatGPT, then as per the guidance I have created the volume using below command: podman volume create -o device=/home/$USER/data myvol the volume got created podman volume inspect myvol { "Name": "myvol", "Driver": "local", "Mountpoint": "/home/noor/.local/share/containers/storage/volumes/myvol/_data", "CreatedAt": "2023-03-01T15:58:43.923688206+05:30", "Labels": {}, "Scope": "local", "Options": { "device": "/home/noor/data" }, "MountCount": 0, "NeedsCopyUp": true, "NeedsChown": true } but when I run the container I get the below error. podman run -it -d --name apache -v myvol:/usr/local/apache2/htdocs -p 8080:80 httpd Error: error mounting volume myvol for container 0715cd8b6a02ac6ff069c2053a40992e4cacfff1d18fad6fa0e12b551dc10335: mount: /home/noor/.local/share/containers/storage/volumes/myvol/_data: permission denied. Any guidance would be helpful.
Create podman volume with custom path I am trying to create a volume with custom path, I want the volume data path to be /home/$USER/data instead of default path which is ~/.local/share/containers/storage/volumes . I have googled it but could not find the result, so I got help from chatGPT, then as per the guidance I have created the volume using below command: podman volume create -o device=/home/$USER/data myvol the volume got created podman volume inspect myvol { "Name": "myvol", "Driver": "local", "Mountpoint": "/home/noor/.local/share/containers/storage/volumes/myvol/_data", "CreatedAt": "2023-03-01T15:58:43.923688206+05:30", "Labels": {}, "Scope": "local", "Options": { "device": "/home/noor/data" }, "MountCount": 0, "NeedsCopyUp": true, "NeedsChown": true } but when I run the container I get the below error. podman run -it -d --name apache -v myvol:/usr/local/apache2/htdocs -p 8080:80 httpd Error: error mounting volume myvol for container 0715cd8b6a02ac6ff069c2053a40992e4cacfff1d18fad6fa0e12b551dc10335: mount: /home/noor/.local/share/containers/storage/volumes/myvol/_data: permission denied. Any guidance would be helpful.
containers, rhel, podman
4
11,150
1
https://stackoverflow.com/questions/75603020/create-podman-volume-with-custom-path
51,838,129
how to change repository location on gitolite?
I recently installed gitolite on my linux server. This server is my git remote server with some bare repositories located on a separate path. lets say /images/git_bare_repos By default gitolite is installed on git user in home directory. I am able to clone gitolite-admin repository from the root user on the same machine. That is gitolite is working fine. >>>ls /home/git bin gitolite projects.list repositories root.pub all the new repositories are created in repositories directory: >>>ls /home/git/repositories/ gitolite-admin.git testing.git My question is: I want gitolite to create new repositories on the same path where all my other bare repositories rest. Ie. in /images/git_bare_repos I tried adding: GL_REPO_BASE => "/images/git_bare_repos", in .gitconfig.rc file but with no luck.
how to change repository location on gitolite? I recently installed gitolite on my linux server. This server is my git remote server with some bare repositories located on a separate path. lets say /images/git_bare_repos By default gitolite is installed on git user in home directory. I am able to clone gitolite-admin repository from the root user on the same machine. That is gitolite is working fine. >>>ls /home/git bin gitolite projects.list repositories root.pub all the new repositories are created in repositories directory: >>>ls /home/git/repositories/ gitolite-admin.git testing.git My question is: I want gitolite to create new repositories on the same path where all my other bare repositories rest. Ie. in /images/git_bare_repos I tried adding: GL_REPO_BASE => "/images/git_bare_repos", in .gitconfig.rc file but with no luck.
linux, git, github, rhel, gitolite
4
1,194
1
https://stackoverflow.com/questions/51838129/how-to-change-repository-location-on-gitolite
33,438,869
Does the mysql client package version on webserver affect PHP queries?
I have two RHEL servers, one to host the PHP application, one to host the MySQL server. Database server has MySQL Enterprise version 5.6.21 installed. While getting the application server built, I asked that the rpm MySQL-client-advanced-5.6.21-1.el6.x86_64 be installed (to match server), but the hardware people don't like this version since 5.6.27 is available which addressed some vulnerabilities. The question is the following: Does the mysql client version on the application server affect the database queries coming from the PHP application? We're using PDO to connect to and query MySQL. If we do this, does the application server even need a mysql client library? Please let me know if I can clarify. Thanks!
Does the mysql client package version on webserver affect PHP queries? I have two RHEL servers, one to host the PHP application, one to host the MySQL server. Database server has MySQL Enterprise version 5.6.21 installed. While getting the application server built, I asked that the rpm MySQL-client-advanced-5.6.21-1.el6.x86_64 be installed (to match server), but the hardware people don't like this version since 5.6.27 is available which addressed some vulnerabilities. The question is the following: Does the mysql client version on the application server affect the database queries coming from the PHP application? We're using PDO to connect to and query MySQL. If we do this, does the application server even need a mysql client library? Please let me know if I can clarify. Thanks!
php, mysql, pdo, rhel
4
420
2
https://stackoverflow.com/questions/33438869/does-the-mysql-client-package-version-on-webserver-affect-php-queries
22,099,571
Convert video to OGV, on fedora platform
I need to develop a rest API for publishing and converting videos. A major hurdle that I am facing, I need a free converter to convert video in any format to OGV. Even though ffmpeg is the best candidate, but I cannot use it because it is not in official Fedora repository, but in rpmfusion. I am developing a product that would be deployed on Fedora infra, they only deploy applications via packages into epel6. Can you guys suggest something from the official fedora repository ?
Convert video to OGV, on fedora platform I need to develop a rest API for publishing and converting videos. A major hurdle that I am facing, I need a free converter to convert video in any format to OGV. Even though ffmpeg is the best candidate, but I cannot use it because it is not in official Fedora repository, but in rpmfusion. I am developing a product that would be deployed on Fedora infra, they only deploy applications via packages into epel6. Can you guys suggest something from the official fedora repository ?
video, ffmpeg, fedora, rhel
4
192
1
https://stackoverflow.com/questions/22099571/convert-video-to-ogv-on-fedora-platform
1,825,414
Use PHP mail to send via smtp
Does anybody know if you can configure php's mail() command so it will only use an SMTP server rather than the local sendmail? We are having trouble with emails being marked as spam. Our server is running RedHat 5 Enterprise. I am aware of various PHP libraries that act as an SMTP client but I'd rather configure PHP so mail() used an SMTP server directly.
Use PHP mail to send via smtp Does anybody know if you can configure php's mail() command so it will only use an SMTP server rather than the local sendmail? We are having trouble with emails being marked as spam. Our server is running RedHat 5 Enterprise. I am aware of various PHP libraries that act as an SMTP client but I'd rather configure PHP so mail() used an SMTP server directly.
php, email, rhel
4
18,177
5
https://stackoverflow.com/questions/1825414/use-php-mail-to-send-via-smtp
61,347,929
Redhat's RHEL8 equivalent of "apk add --no-cache gcc musl-dev linux-headers"
I am trying this sample code from "[URL] on Python installed on RedHat enterprise linux 7 / 8 image. Getting error 127 on this line RUN apk add --no-cache git gcc musl-dev linux-headers Thanks
Redhat's RHEL8 equivalent of "apk add --no-cache gcc musl-dev linux-headers" I am trying this sample code from "[URL] on Python installed on RedHat enterprise linux 7 / 8 image. Getting error 127 on this line RUN apk add --no-cache git gcc musl-dev linux-headers Thanks
rhel, redhat-containers, rhel8
4
4,437
1
https://stackoverflow.com/questions/61347929/redhats-rhel8-equivalent-of-apk-add-no-cache-gcc-musl-dev-linux-headers
59,162,343
Docker run - apk update - could not connect to server
I'm trying to run a simple alpine docker image on rhel 7, but i'm stuck in this error when i run apk update: fetch [URL] ERROR: [URL] could not connect to server (check repositories file) WARNING: Ignoring APKINDEX.00740ba1.tar.gz: No such file or directory fetch [URL] ERROR: [URL] could not connect to server (check repositories file) WARNING: Ignoring APKINDEX.d8b2a6f4.tar.gz: No such file or directory 2 errors; 14 distinct packages available I run these commands: docker run -it --rm --env="http_proxy=[URL] alpine /bin/sh apk update I'm behind a ssh tunnel proxy in the virtual machine. I've set /etc/docker/daemon_config.json, /etc/systemd/system/docker.service.d/http.proxy.d, enviroment variables and dns. What can i do?
Docker run - apk update - could not connect to server I'm trying to run a simple alpine docker image on rhel 7, but i'm stuck in this error when i run apk update: fetch [URL] ERROR: [URL] could not connect to server (check repositories file) WARNING: Ignoring APKINDEX.00740ba1.tar.gz: No such file or directory fetch [URL] ERROR: [URL] could not connect to server (check repositories file) WARNING: Ignoring APKINDEX.d8b2a6f4.tar.gz: No such file or directory 2 errors; 14 distinct packages available I run these commands: docker run -it --rm --env="http_proxy=[URL] alpine /bin/sh apk update I'm behind a ssh tunnel proxy in the virtual machine. I've set /etc/docker/daemon_config.json, /etc/systemd/system/docker.service.d/http.proxy.d, enviroment variables and dns. What can i do?
docker, proxy, dns, docker-compose, rhel
4
5,352
0
https://stackoverflow.com/questions/59162343/docker-run-apk-update-could-not-connect-to-server
41,593,349
Tomcat 8 org.apache.commons.io.FilenameUtil
I have a web application that uses Apache Common IO version 2.4. When I deploy this application on Tomcat 8.0.39 I am getting following error: > [ERROR] - class "org.apache.commons.io.FilenameUtils"'s signer information does not match signer information of other classes in the same package java.lang.SecurityException: class "org.apache.commons.io.FilenameUtils"'s signer information does not match signer information of other classes in the same package at java.lang.ClassLoader.checkCerts(ClassLoader.java:898) at java.lang.ClassLoader.preDefineClass(ClassLoader.java:668) at java.lang.ClassLoader.defineClass(ClassLoader.java:761) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) at org.apache.catalina.loader.WebappClassLoaderBase.findClassInternal(WebappClassLoaderBase.java:2558) at org.apache.catalina.loader.WebappClassLoaderBase.findClass(WebappClassLoaderBase.java:859) at org.apache.catalina.loader.WebappClassLoaderBase.loadClass(WebappClassLoaderBase.java:1302) at org.apache.catalina.loader.WebappClassLoaderBase.loadClass(WebappClassLoaderBase.java:1167) at com.b.aD.a(Unknown Source) at com.b.Bm.getDataSet(Unknown Source) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.camel.component.bean.MethodInfo.invoke(MethodInfo.java:408) at org.apache.camel.component.bean.MethodInfo$1.doProceed(MethodInfo.java:279) at org.apache.camel.component.bean.MethodInfo$1.proceed(MethodInfo.java:252) at org.apache.camel.component.bean.BeanProcessor.process(BeanProcessor.java:171) at org.apache.camel.util.AsyncProcessorHelper.process(AsyncProcessorHelper.java:109) at org.apache.camel.component.bean.BeanProcessor.process(BeanProcessor.java:68) at org.apache.camel.component.bean.BeanProducer.process(BeanProducer.java:38) at org.apache.camel.processor.SendProcessor.process(SendProcessor.java:129) This issue particularly faced in RHEL Box not in windows
Tomcat 8 org.apache.commons.io.FilenameUtil I have a web application that uses Apache Common IO version 2.4. When I deploy this application on Tomcat 8.0.39 I am getting following error: > [ERROR] - class "org.apache.commons.io.FilenameUtils"'s signer information does not match signer information of other classes in the same package java.lang.SecurityException: class "org.apache.commons.io.FilenameUtils"'s signer information does not match signer information of other classes in the same package at java.lang.ClassLoader.checkCerts(ClassLoader.java:898) at java.lang.ClassLoader.preDefineClass(ClassLoader.java:668) at java.lang.ClassLoader.defineClass(ClassLoader.java:761) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) at org.apache.catalina.loader.WebappClassLoaderBase.findClassInternal(WebappClassLoaderBase.java:2558) at org.apache.catalina.loader.WebappClassLoaderBase.findClass(WebappClassLoaderBase.java:859) at org.apache.catalina.loader.WebappClassLoaderBase.loadClass(WebappClassLoaderBase.java:1302) at org.apache.catalina.loader.WebappClassLoaderBase.loadClass(WebappClassLoaderBase.java:1167) at com.b.aD.a(Unknown Source) at com.b.Bm.getDataSet(Unknown Source) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.camel.component.bean.MethodInfo.invoke(MethodInfo.java:408) at org.apache.camel.component.bean.MethodInfo$1.doProceed(MethodInfo.java:279) at org.apache.camel.component.bean.MethodInfo$1.proceed(MethodInfo.java:252) at org.apache.camel.component.bean.BeanProcessor.process(BeanProcessor.java:171) at org.apache.camel.util.AsyncProcessorHelper.process(AsyncProcessorHelper.java:109) at org.apache.camel.component.bean.BeanProcessor.process(BeanProcessor.java:68) at org.apache.camel.component.bean.BeanProducer.process(BeanProducer.java:38) at org.apache.camel.processor.SendProcessor.process(SendProcessor.java:129) This issue particularly faced in RHEL Box not in windows
java, apache, tomcat, rhel, rhel6
4
659
0
https://stackoverflow.com/questions/41593349/tomcat-8-org-apache-commons-io-filenameutil
39,184,465
Passing argument containg nullbyte (\x00) to subprocess.call
#!/usr/bin/env python from subprocess import call try: call(['echo', "aabbccddee".decode('hex')]) print 'Input without \\x00 works!' except Exception as e: print 'Input without \\x00 throws exception: ' + str(e) try: call(['echo', "aabbccdd00ee".decode('hex')]) # The input contains \x00 print 'Input with \\x00 works!' except Exception as e: print 'Input with \\x00 throws exception: ' + str(e) Returns the following (tested with Python 2.7 on RHEL 6): ▒▒▒▒▒ Input without \x00 works! Input with \x00 throws exception: execv() arg 2 must contain only strings Questions: Is this expected behavior by subprocess.call or might this be a bug? Is there any way I can pass binary data (directly) containing \x00 to another executable within a python script? It will probably work using shell=True, but I would prefer another way, if there is any. Note: >>> type("00".decode('hex')) <type 'str'> Running echo $'\x00' directly works as expected. EDIT This seems to be correct behavior after more research. Due to the execve(2) semantics it is not possible to pass a string containing a null byte as argument. See: $ echo $'\x55\x55\x55\x00\x55\x55' UUU However, I couldn't really find anymore info on this (second question especially, as it seems shell=True won't help either (all data after the null byte wouldn't be passed)). I leave this question open for a while, maybe someone can provide a deeper insight or share some useful resources on this.
Passing argument containg nullbyte (\x00) to subprocess.call #!/usr/bin/env python from subprocess import call try: call(['echo', "aabbccddee".decode('hex')]) print 'Input without \\x00 works!' except Exception as e: print 'Input without \\x00 throws exception: ' + str(e) try: call(['echo', "aabbccdd00ee".decode('hex')]) # The input contains \x00 print 'Input with \\x00 works!' except Exception as e: print 'Input with \\x00 throws exception: ' + str(e) Returns the following (tested with Python 2.7 on RHEL 6): ▒▒▒▒▒ Input without \x00 works! Input with \x00 throws exception: execv() arg 2 must contain only strings Questions: Is this expected behavior by subprocess.call or might this be a bug? Is there any way I can pass binary data (directly) containing \x00 to another executable within a python script? It will probably work using shell=True, but I would prefer another way, if there is any. Note: >>> type("00".decode('hex')) <type 'str'> Running echo $'\x00' directly works as expected. EDIT This seems to be correct behavior after more research. Due to the execve(2) semantics it is not possible to pass a string containing a null byte as argument. See: $ echo $'\x55\x55\x55\x00\x55\x55' UUU However, I couldn't really find anymore info on this (second question especially, as it seems shell=True won't help either (all data after the null byte wouldn't be passed)). I leave this question open for a while, maybe someone can provide a deeper insight or share some useful resources on this.
linux, python-2.7, subprocess, rhel
4
1,195
0
https://stackoverflow.com/questions/39184465/passing-argument-containg-nullbyte-x00-to-subprocess-call
27,353,951
Error: no display specified
I am trying to open firefox browser on redhat server using firefox command but i get following error message. firefox Error: no display specified How to fix this error?
Error: no display specified I am trying to open firefox browser on redhat server using firefox command but i get following error message. firefox Error: no display specified How to fix this error?
firefox, rhel
4
12,009
0
https://stackoverflow.com/questions/27353951/error-no-display-specified
63,934,718
Gitlab Runner: microdnf/yum command not found on different machine
I have two different server machine with gitlab-runner hosted. I use one as development version, the other one as production release. I have the same configuration on both machine, so the lastest release of gitlab-runner and the lastest release of Docker/docker-compose. Into my dockerfiles I need to download net-tools package in order to use netstat command. The builder docker is an official image of OpenJDK 14 and it is RHEL based. So, in order to install a new package, i can use only dnf/microdnf or yum. Ok, everything should work simple and clear as exposed so far. But... when the machines try to run the command: RUN microdnf install net-tools On staging machine: /bin/sh: microdnf: command not found Instead, on production machine: Job succeeded. Okay, so what about yum? Let's change Dockerfile's net-tools installation RUN yum install net-tools -y Pipeline started, job scheduled and... staging machine: Job succeeded. On production machine, as you can figure: /bin/sh: yum: command not found I feel i little bit trapped because there is not another way to install package (dnf should be another possible way but it is not install on both docker images) and I don't want apply a workaround that can "test" the installation via microdnf or yum. I hope I've made the problem as clear as I can.
Gitlab Runner: microdnf/yum command not found on different machine I have two different server machine with gitlab-runner hosted. I use one as development version, the other one as production release. I have the same configuration on both machine, so the lastest release of gitlab-runner and the lastest release of Docker/docker-compose. Into my dockerfiles I need to download net-tools package in order to use netstat command. The builder docker is an official image of OpenJDK 14 and it is RHEL based. So, in order to install a new package, i can use only dnf/microdnf or yum. Ok, everything should work simple and clear as exposed so far. But... when the machines try to run the command: RUN microdnf install net-tools On staging machine: /bin/sh: microdnf: command not found Instead, on production machine: Job succeeded. Okay, so what about yum? Let's change Dockerfile's net-tools installation RUN yum install net-tools -y Pipeline started, job scheduled and... staging machine: Job succeeded. On production machine, as you can figure: /bin/sh: yum: command not found I feel i little bit trapped because there is not another way to install package (dnf should be another possible way but it is not install on both docker images) and I don't want apply a workaround that can "test" the installation via microdnf or yum. I hope I've made the problem as clear as I can.
docker, gitlab-ci-runner, yum, rhel, dnf
4
8,293
1
https://stackoverflow.com/questions/63934718/gitlab-runner-microdnf-yum-command-not-found-on-different-machine
45,744,883
Wait functionality in Wget not working as expected
Background: I am using wget (version 1.14-15.el7) on RedHat Linux (version 7.3 (Maipo)), in a Bash (version 4.2.46(1)) terminal. Unfortunately, I am limited to this particular RHEL image, so I can't upgrade to a newer version of wget. Goal: I am trying to configure wget so that, if a download attempt fails, it does the following: 1) Retries the download 4 more times (total of 5 attempts) 2) Waits a FIXED amount of time (30 seconds) between download attempts ('retries') For context, here is a relevant snippet of the wget manual: -w seconds --wait=seconds Wait the specified number of seconds between the retrievals. Use of this option is recommended, as it lightens the server load by making the requests less frequent. Instead of in seconds, the time can be specified in minutes using the "m" suffix, in hours using "h" suffix, or in days using "d" suffix. Specifying a large value for this option is useful if the network or the destination host is down, so that Wget can wait long enough to reasonably expect the network error to be fixed before the retry. The waiting interval specified by this function is influenced by "--random-wait", which see. --waitretry=seconds If you don't want Wget to wait between every retrieval, but only between retries of failed downloads, you can use this option. Wget will use linear backoff, waiting 1 second after the first failure on a given file, then waiting 2 seconds after the second failure on that file, up to the maximum number of seconds you specify. By default, Wget will assume a value of 10 seconds. To be clear, I am using the --wait flag, NOT the --waitretry flag. Process: First, I export/set a wrong http_proxy and https_proxy, to ensure any download attempt will time out. I run the following command: wget --tries=5 --wait=30 <url> -O <output_filename> At this point, the --wait functionality does NOT work as expected. Specifically, it does NOT wait 30s after each download attempt. Instead: 1) After the first attempt, it waits 1s. 2) After the 2nd attempt, it waits 2s. 3) After the 3rd attempt, it waits 3s. and so on ... In other words, despite using the --wait flag (which should result in a fixed waiting time between download attempts), wget seems to be performing a 'linear backoff' as described in the --waitretry flag section. PROBLEM: I want the functionality of the --wait flag, NOT the --waitretry flag. Unfortunately, the --wait flag seems to be acting like the --waitretry flag -- is there any way to get around this apparent bug in wget, so that using the --wait flag results in the expected fixed waiting time between download attempts?
Wait functionality in Wget not working as expected Background: I am using wget (version 1.14-15.el7) on RedHat Linux (version 7.3 (Maipo)), in a Bash (version 4.2.46(1)) terminal. Unfortunately, I am limited to this particular RHEL image, so I can't upgrade to a newer version of wget. Goal: I am trying to configure wget so that, if a download attempt fails, it does the following: 1) Retries the download 4 more times (total of 5 attempts) 2) Waits a FIXED amount of time (30 seconds) between download attempts ('retries') For context, here is a relevant snippet of the wget manual: -w seconds --wait=seconds Wait the specified number of seconds between the retrievals. Use of this option is recommended, as it lightens the server load by making the requests less frequent. Instead of in seconds, the time can be specified in minutes using the "m" suffix, in hours using "h" suffix, or in days using "d" suffix. Specifying a large value for this option is useful if the network or the destination host is down, so that Wget can wait long enough to reasonably expect the network error to be fixed before the retry. The waiting interval specified by this function is influenced by "--random-wait", which see. --waitretry=seconds If you don't want Wget to wait between every retrieval, but only between retries of failed downloads, you can use this option. Wget will use linear backoff, waiting 1 second after the first failure on a given file, then waiting 2 seconds after the second failure on that file, up to the maximum number of seconds you specify. By default, Wget will assume a value of 10 seconds. To be clear, I am using the --wait flag, NOT the --waitretry flag. Process: First, I export/set a wrong http_proxy and https_proxy, to ensure any download attempt will time out. I run the following command: wget --tries=5 --wait=30 <url> -O <output_filename> At this point, the --wait functionality does NOT work as expected. Specifically, it does NOT wait 30s after each download attempt. Instead: 1) After the first attempt, it waits 1s. 2) After the 2nd attempt, it waits 2s. 3) After the 3rd attempt, it waits 3s. and so on ... In other words, despite using the --wait flag (which should result in a fixed waiting time between download attempts), wget seems to be performing a 'linear backoff' as described in the --waitretry flag section. PROBLEM: I want the functionality of the --wait flag, NOT the --waitretry flag. Unfortunately, the --wait flag seems to be acting like the --waitretry flag -- is there any way to get around this apparent bug in wget, so that using the --wait flag results in the expected fixed waiting time between download attempts?
linux, wget, rhel
4
1,973
1
https://stackoverflow.com/questions/45744883/wait-functionality-in-wget-not-working-as-expected
52,990,155
I&#39;m getting Error when install Docker on RHEL
I'm newbie on Linux and Docker platform and trying to install Docker on RHEL by repository method as stated in Docker installation guide but when I type sudo yum -y install docker-ee I'm getting that output and I cannot continue to installation. I think that the previous steps performed correctly but can not go on other steps due to this output : [root@localhost yum.repos.d]# sudo yum -y install docker-ee Loaded plugins: langpacks, product-id, search-disabled-repos, : subscription-manager rhel-7-server-extras-rpms | 3.4 kB 00:00 rhel-7-server-rpms | 3.5 kB 00:00 (1/3): rhel-7-server-extras-rpms/x86_64/group | 104 B 00:00 (2/3): rhel-7-server-extras-rpms/x86_64/updatei | 280 kB 00:00 (3/3): rhel-7-server-extras-rpms/x86_64/primary | 446 kB 00:00 No package docker-ee available. Error: Nothing to do Previous steps by Docker guide : sudo rm /etc/yum.repos.d/docker*.repo export export DOCKERURL="<DOCKER-EE-URL>" sudo -E sh -c 'echo "$DOCKERURL/rhel" > /etc/yum/vars/dockerurl' sudo sh -c 'echo "7" > /etc/yum/vars/dockerosversion' sudo yum install -y yum-utils \ device-mapper-persistent-data \lvm2 sudo yum-config-manager --enable rhel-7-server-extras-rpms sudo -E yum-config-manager \--add-repo \"$DOCKERURL/rhel/docker-ee.repo" sudo yum-config-manager --enable docker-ee-stable-18.03 Current step : sudo yum -y install docker-ee What can be the problem ? Thank you so much
I&#39;m getting Error when install Docker on RHEL I'm newbie on Linux and Docker platform and trying to install Docker on RHEL by repository method as stated in Docker installation guide but when I type sudo yum -y install docker-ee I'm getting that output and I cannot continue to installation. I think that the previous steps performed correctly but can not go on other steps due to this output : [root@localhost yum.repos.d]# sudo yum -y install docker-ee Loaded plugins: langpacks, product-id, search-disabled-repos, : subscription-manager rhel-7-server-extras-rpms | 3.4 kB 00:00 rhel-7-server-rpms | 3.5 kB 00:00 (1/3): rhel-7-server-extras-rpms/x86_64/group | 104 B 00:00 (2/3): rhel-7-server-extras-rpms/x86_64/updatei | 280 kB 00:00 (3/3): rhel-7-server-extras-rpms/x86_64/primary | 446 kB 00:00 No package docker-ee available. Error: Nothing to do Previous steps by Docker guide : sudo rm /etc/yum.repos.d/docker*.repo export export DOCKERURL="<DOCKER-EE-URL>" sudo -E sh -c 'echo "$DOCKERURL/rhel" > /etc/yum/vars/dockerurl' sudo sh -c 'echo "7" > /etc/yum/vars/dockerosversion' sudo yum install -y yum-utils \ device-mapper-persistent-data \lvm2 sudo yum-config-manager --enable rhel-7-server-extras-rpms sudo -E yum-config-manager \--add-repo \"$DOCKERURL/rhel/docker-ee.repo" sudo yum-config-manager --enable docker-ee-stable-18.03 Current step : sudo yum -y install docker-ee What can be the problem ? Thank you so much
linux, docker, containers, rhel
3
8,725
2
https://stackoverflow.com/questions/52990155/im-getting-error-when-install-docker-on-rhel
68,346,692
Failed to start MariaDB 10.3 database server
Unable to start mariadb.service. I'm getting following error while starting mariadb.service. [root@localhost bggb]# systemctl start mariadb.service Job for mariadb.service failed because the control process exited with error code. See "systemctl status mariadb.service" and "journalctl -xe" for details. Here is the status of mariadb.service [root@localhost bggb]# systemctl status mariadb.service ● mariadb.service - MariaDB 10.3 database server Loaded: loaded (/usr/lib/systemd/system/mariadb.service; enabled; vendor preset: disabled) Active: failed (Result: exit-code) since Mon 2021-07-12 16:56:03 IST; 59s ago Docs: man:mysqld(8) [URL] Process: 6260 ExecStart=/usr/libexec/mysqld --basedir=/usr $MYSQLD_OPTS $_WSREP_NEW_CLUSTER (code=exited> Process: 6221 ExecStartPre=/usr/libexec/mysql-prepare-db-dir mariadb.service (code=exited, status=0/SUCC> Process: 6196 ExecStartPre=/usr/libexec/mysql-check-socket (code=exited, status=0/SUCCESS) Main PID: 6260 (code=exited, status=1/FAILURE) Status: "MariaDB server is down" Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [ERROR] Plugin 'Aria' init funct> Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [ERROR] Plugin 'Aria' registrati> Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [Note] Plugin 'FEEDBACK' is disa> Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [ERROR] Could not open mysql.plu> Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [ERROR] Unknown/unsupported stor> Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [ERROR] Aborting Jul 12 16:56:03 localhost.localdomain mysqld[6260]: Warning: Memory not freed: 520 Jul 12 16:56:03 localhost.localdomain systemd[1]: mariadb.service: Main process exited, code=exited, statu> Jul 12 16:56:03 localhost.localdomain systemd[1]: mariadb.service: Failed with result 'exit-code'. Jul 12 16:56:03 localhost.localdomain systemd[1]: Failed to start MariaDB 10.3 database server. Hear is the output of journalctl -xe for more information as requested. Configuration is done first time. [root@localhost bggb]# journalctl -xe Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: == Stack trace for context 0x55df8c69> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #0 7ffd7797b5d0 b resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #1 7ffd7797b680 b resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #2 55df8cc77d30 i resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #3 7ffd7797c610 b resource:///org> Jul 12 17:45:52 localhost.localdomain gnome-shell[2874]: Object .Gjs_CapsLockWarning (0x55df8d192040), has> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #4 55df8cc77ca0 i resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #5 55df8cc77c20 i resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: == Stack trace for context 0x55df8c69> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #0 7ffd7797b5d0 b resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #1 7ffd7797b680 b resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #2 55df8cc77d30 i resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #3 7ffd7797c610 b resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #4 55df8cc77ca0 i resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #5 55df8cc77c20 i resource:///org> Jul 12 17:45:52 localhost.localdomain gnome-shell[2874]: Object .Gjs_CapsLockWarning (0x55df8d192040), has> Jul 12 17:45:53 localhost.localdomain NetworkManager[1115]: <info> [1626092153.0272] agent-manager: agent> Jul 12 17:46:03 localhost.localdomain systemd[1]: fprintd.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: [URL] -- -- The unit fprintd.service has successfully entered the 'dead' state.
Failed to start MariaDB 10.3 database server Unable to start mariadb.service. I'm getting following error while starting mariadb.service. [root@localhost bggb]# systemctl start mariadb.service Job for mariadb.service failed because the control process exited with error code. See "systemctl status mariadb.service" and "journalctl -xe" for details. Here is the status of mariadb.service [root@localhost bggb]# systemctl status mariadb.service ● mariadb.service - MariaDB 10.3 database server Loaded: loaded (/usr/lib/systemd/system/mariadb.service; enabled; vendor preset: disabled) Active: failed (Result: exit-code) since Mon 2021-07-12 16:56:03 IST; 59s ago Docs: man:mysqld(8) [URL] Process: 6260 ExecStart=/usr/libexec/mysqld --basedir=/usr $MYSQLD_OPTS $_WSREP_NEW_CLUSTER (code=exited> Process: 6221 ExecStartPre=/usr/libexec/mysql-prepare-db-dir mariadb.service (code=exited, status=0/SUCC> Process: 6196 ExecStartPre=/usr/libexec/mysql-check-socket (code=exited, status=0/SUCCESS) Main PID: 6260 (code=exited, status=1/FAILURE) Status: "MariaDB server is down" Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [ERROR] Plugin 'Aria' init funct> Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [ERROR] Plugin 'Aria' registrati> Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [Note] Plugin 'FEEDBACK' is disa> Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [ERROR] Could not open mysql.plu> Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [ERROR] Unknown/unsupported stor> Jul 12 16:56:03 localhost.localdomain mysqld[6260]: 2021-07-12 16:56:03 0 [ERROR] Aborting Jul 12 16:56:03 localhost.localdomain mysqld[6260]: Warning: Memory not freed: 520 Jul 12 16:56:03 localhost.localdomain systemd[1]: mariadb.service: Main process exited, code=exited, statu> Jul 12 16:56:03 localhost.localdomain systemd[1]: mariadb.service: Failed with result 'exit-code'. Jul 12 16:56:03 localhost.localdomain systemd[1]: Failed to start MariaDB 10.3 database server. Hear is the output of journalctl -xe for more information as requested. Configuration is done first time. [root@localhost bggb]# journalctl -xe Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: == Stack trace for context 0x55df8c69> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #0 7ffd7797b5d0 b resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #1 7ffd7797b680 b resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #2 55df8cc77d30 i resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #3 7ffd7797c610 b resource:///org> Jul 12 17:45:52 localhost.localdomain gnome-shell[2874]: Object .Gjs_CapsLockWarning (0x55df8d192040), has> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #4 55df8cc77ca0 i resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #5 55df8cc77c20 i resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: == Stack trace for context 0x55df8c69> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #0 7ffd7797b5d0 b resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #1 7ffd7797b680 b resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #2 55df8cc77d30 i resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #3 7ffd7797c610 b resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #4 55df8cc77ca0 i resource:///org> Jul 12 17:45:52 localhost.localdomain org.gnome.Shell.desktop[2874]: #5 55df8cc77c20 i resource:///org> Jul 12 17:45:52 localhost.localdomain gnome-shell[2874]: Object .Gjs_CapsLockWarning (0x55df8d192040), has> Jul 12 17:45:53 localhost.localdomain NetworkManager[1115]: <info> [1626092153.0272] agent-manager: agent> Jul 12 17:46:03 localhost.localdomain systemd[1]: fprintd.service: Succeeded. -- Subject: Unit succeeded -- Defined-By: systemd -- Support: [URL] -- -- The unit fprintd.service has successfully entered the 'dead' state.
mariadb, rhel
3
46,052
2
https://stackoverflow.com/questions/68346692/failed-to-start-mariadb-10-3-database-server
58,113,653
Using Ansible loop to create multiple users: Undefined error
I am using the following ansible code to create multiple unix user accounts --- - hosts: test become: true tasks: - name: more complex items to add several users user: name: "{{ item.name }}" uid: "{{ item.uid }}" groups: "{{ item.groups }}" state: present with_items: "{{ user_details }}" I am storing the user information by using a separate a variable file as below cat /etc/ansible/vars.yml --- user_details: - { name: testuser1, uid: 1002, groups: "admin, logs" } - { name: testuser2, uid: 1003, groups: logs: } To execute above playbook , I tried with both the commands below sudo ansible-playbook /etc/ansible/userloop.yml -e /etc/ansible/vars.yml sudo ansible-playbook /etc/ansible/userloop.yml but both commands are failing with below error fatal: [host-003]: FAILED! => {"msg": "'user_details' is undefined"} fatal: [host-004]: FAILED! => {"msg": "'user_details' is undefined"} How to resolve the issue ? I want to maintain a separate variable file to store the user information rather then putting them in the same playbook file .
Using Ansible loop to create multiple users: Undefined error I am using the following ansible code to create multiple unix user accounts --- - hosts: test become: true tasks: - name: more complex items to add several users user: name: "{{ item.name }}" uid: "{{ item.uid }}" groups: "{{ item.groups }}" state: present with_items: "{{ user_details }}" I am storing the user information by using a separate a variable file as below cat /etc/ansible/vars.yml --- user_details: - { name: testuser1, uid: 1002, groups: "admin, logs" } - { name: testuser2, uid: 1003, groups: logs: } To execute above playbook , I tried with both the commands below sudo ansible-playbook /etc/ansible/userloop.yml -e /etc/ansible/vars.yml sudo ansible-playbook /etc/ansible/userloop.yml but both commands are failing with below error fatal: [host-003]: FAILED! => {"msg": "'user_details' is undefined"} fatal: [host-004]: FAILED! => {"msg": "'user_details' is undefined"} How to resolve the issue ? I want to maintain a separate variable file to store the user information rather then putting them in the same playbook file .
ansible, rhel, user-accounts
3
18,587
3
https://stackoverflow.com/questions/58113653/using-ansible-loop-to-create-multiple-users-undefined-error
65,808,589
Issue with Docker installation on Linux
I am trying to install Docker on RHEL whose version details are shown below I have downloaded the rpm docker-ce-20.10.2-3.el7.x86_64.rpm from here . And I run the following install command sudo yum install /home/projuser/usr/share/Docker/docker-ce-20.10.2-3.el7.x86_64.rpm Error comes up suggesting that docker-ce-rootless-extras is required. Error: Package: 3:docker-ce-20.10.2-3.el7.x86_64 (/docker-ce-20.10.2-3.el7.x86_64) Requires: docker-ce-rootless-extras Here is the full screenshot of error As per the message I try to install docker-ce-rootless-extras-20.10.2-3.el7.x86_64.rpm that I downloaded from here using command shown below sudo yum install /home/projuser/usr/share/Docker/docker-ce-rootless-extras-20.10.2-3.el7.x86_64.rpm I get an error that suggests that docker-ce is required Error: Package: docker-ce-rootless-extras-20.10.2-3.el7.x86_64 (/docker-ce-rootless-extras-20.10.2-3.el7.x86_64) Requires: docker-ce Here is the complete screenshot So I am not sure how to resolve this dependency as docker-ce and docker-ce-rootless-extras are both not getting installed with dependency on each other. Please let me know if you need any other details. Please help resolve this
Issue with Docker installation on Linux I am trying to install Docker on RHEL whose version details are shown below I have downloaded the rpm docker-ce-20.10.2-3.el7.x86_64.rpm from here . And I run the following install command sudo yum install /home/projuser/usr/share/Docker/docker-ce-20.10.2-3.el7.x86_64.rpm Error comes up suggesting that docker-ce-rootless-extras is required. Error: Package: 3:docker-ce-20.10.2-3.el7.x86_64 (/docker-ce-20.10.2-3.el7.x86_64) Requires: docker-ce-rootless-extras Here is the full screenshot of error As per the message I try to install docker-ce-rootless-extras-20.10.2-3.el7.x86_64.rpm that I downloaded from here using command shown below sudo yum install /home/projuser/usr/share/Docker/docker-ce-rootless-extras-20.10.2-3.el7.x86_64.rpm I get an error that suggests that docker-ce is required Error: Package: docker-ce-rootless-extras-20.10.2-3.el7.x86_64 (/docker-ce-rootless-extras-20.10.2-3.el7.x86_64) Requires: docker-ce Here is the complete screenshot So I am not sure how to resolve this dependency as docker-ce and docker-ce-rootless-extras are both not getting installed with dependency on each other. Please let me know if you need any other details. Please help resolve this
linux, docker, rhel
3
6,217
3
https://stackoverflow.com/questions/65808589/issue-with-docker-installation-on-linux
74,548,251
Print Date and Username from a secure log file
I'm trying to get the Username and Date 'y/m/d' from a /var/log/secure , however when I'm trying to use awk , it only provides the date. Sample of a secure log file. 2022-11-23T02:03:24.594880+01:00 servername su: pam_unix(su:session): session opened for user john.doe by (uid=0) What I'm expecting to print is: 2022-11-23 john.doe Here's my code. cat /var/log/secure | grep 'session opened' | awk -FT '{print $1 " " " User: " $9 }' "The output is only: 2022-11-23 User:"
Print Date and Username from a secure log file I'm trying to get the Username and Date 'y/m/d' from a /var/log/secure , however when I'm trying to use awk , it only provides the date. Sample of a secure log file. 2022-11-23T02:03:24.594880+01:00 servername su: pam_unix(su:session): session opened for user john.doe by (uid=0) What I'm expecting to print is: 2022-11-23 john.doe Here's my code. cat /var/log/secure | grep 'session opened' | awk -FT '{print $1 " " " User: " $9 }' "The output is only: 2022-11-23 User:"
bash, awk, rhel
3
372
4
https://stackoverflow.com/questions/74548251/print-date-and-username-from-a-secure-log-file
27,409,037
Python 2.7.6 error when building from source: gcc.orig: directory&quot;: No such file or directory
I've been provisioned a very bare-bones RHEL 4.4 at work; e.g., it didn't have java or gcc installed. I have followed this guide to install python 2.7.6 on another VM (RHEM 4.6) successfully. However on this new vm, I cannot make without getting this error: gcc -pthread -c -fno-strict-aliasing -g -O2 -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -I. -IInclude -I./Include -DPy_BUILD_CORE \ -DSVNVERSION="\"LC_ALL=C echo Unversioned directory\"" \ -DHGVERSION="\"LC_ALL=C \"" \ -DHGTAG="\"LC_ALL=C \"" \ -DHGBRANCH="\"LC_ALL=C \"" \ -o Modules/getbuildinfo.o ./Modules/getbuildinfo.c gcc.orig: directory": No such file or directory <command-line>: warning: missing terminating " character ./Modules/getbuildinfo.c: In function â_Py_svnversionâ: ./Modules/getbuildinfo.c:63: error: missing terminating " character ./Modules/getbuildinfo.c:63: error: expected expression before â;â token make: *** [Modules/getbuildinfo.o] Error 1 I tried what was suggested in this unrelated bug report: SVNVERSION="Unversioned directory" ./configure make but received the same error. I saw this question on super user, which suggests to check the output of the svnversion command in the Modules directory, but I don't have svn or svnversion on this machine. It also suggests to edit the function _PY_svnversion in Modules/getbuildinfo.c , which I did but received the same error. Any ideas?
Python 2.7.6 error when building from source: gcc.orig: directory&quot;: No such file or directory I've been provisioned a very bare-bones RHEL 4.4 at work; e.g., it didn't have java or gcc installed. I have followed this guide to install python 2.7.6 on another VM (RHEM 4.6) successfully. However on this new vm, I cannot make without getting this error: gcc -pthread -c -fno-strict-aliasing -g -O2 -DNDEBUG -g -fwrapv -O3 -Wall -Wstrict-prototypes -I. -IInclude -I./Include -DPy_BUILD_CORE \ -DSVNVERSION="\"LC_ALL=C echo Unversioned directory\"" \ -DHGVERSION="\"LC_ALL=C \"" \ -DHGTAG="\"LC_ALL=C \"" \ -DHGBRANCH="\"LC_ALL=C \"" \ -o Modules/getbuildinfo.o ./Modules/getbuildinfo.c gcc.orig: directory": No such file or directory <command-line>: warning: missing terminating " character ./Modules/getbuildinfo.c: In function â_Py_svnversionâ: ./Modules/getbuildinfo.c:63: error: missing terminating " character ./Modules/getbuildinfo.c:63: error: expected expression before â;â token make: *** [Modules/getbuildinfo.o] Error 1 I tried what was suggested in this unrelated bug report: SVNVERSION="Unversioned directory" ./configure make but received the same error. I saw this question on super user, which suggests to check the output of the svnversion command in the Modules directory, but I don't have svn or svnversion on this machine. It also suggests to edit the function _PY_svnversion in Modules/getbuildinfo.c , which I did but received the same error. Any ideas?
python, python-2.7, rhel
3
1,352
1
https://stackoverflow.com/questions/27409037/python-2-7-6-error-when-building-from-source-gcc-orig-directory-no-such-file
54,345,009
pam_open_session: System error and sudo: policy plugin failed session initialization with no pass expiary in etc/pam.d/password-auth
I am struggeling with an error in REHL 7. I have a REHL Image witch is accessible only over SSH key without user/password credentials. To prevent the password renew process after 90 days (which is not possible with SSH key) I added the no_pass_expiry in my etc/pam.d/password-auth but then i receive the following error when i try to sudo pam.d]$ sudo su - sudo: pam_open_session: System error sudo: policy plugin failed session initialization this only occures after the 90 day.
pam_open_session: System error and sudo: policy plugin failed session initialization with no pass expiary in etc/pam.d/password-auth I am struggeling with an error in REHL 7. I have a REHL Image witch is accessible only over SSH key without user/password credentials. To prevent the password renew process after 90 days (which is not possible with SSH key) I added the no_pass_expiry in my etc/pam.d/password-auth but then i receive the following error when i try to sudo pam.d]$ sudo su - sudo: pam_open_session: System error sudo: policy plugin failed session initialization this only occures after the 90 day.
rhel, rhel7, pam
3
25,561
2
https://stackoverflow.com/questions/54345009/pam-open-session-system-error-and-sudo-policy-plugin-failed-session-initializa
51,818,610
.Net Core application exit code. Different behaviour on Linux (RHEL 7)
I seem unable to set the exit code of a dotnet core console application such that the expected behaviour is observed on RHEL7. I am using dotnet 2.0.3. using System; namespace HelloWorld { class Program { static void Main(string[] args) { Console.WriteLine("Hello World!"); Environment.Exit(12345); } } } On windows, the error level is set correctly when executing the program with dotnet HelloWorld.dll echo %errorlevel% 12345 On RHEL7, this is not what is seen dotnet HelloWorld.dll echo $? 57 I have tried various ways to return an exit code (setting the Environment.ExitCode property, returning an integer from the program) but have had no success. I will try a test on the latest dotnet core version (2.1.302) as soon as possible but posting this question in the meantime. Does anyone know what's going on here? Thanks
.Net Core application exit code. Different behaviour on Linux (RHEL 7) I seem unable to set the exit code of a dotnet core console application such that the expected behaviour is observed on RHEL7. I am using dotnet 2.0.3. using System; namespace HelloWorld { class Program { static void Main(string[] args) { Console.WriteLine("Hello World!"); Environment.Exit(12345); } } } On windows, the error level is set correctly when executing the program with dotnet HelloWorld.dll echo %errorlevel% 12345 On RHEL7, this is not what is seen dotnet HelloWorld.dll echo $? 57 I have tried various ways to return an exit code (setting the Environment.ExitCode property, returning an integer from the program) but have had no success. I will try a test on the latest dotnet core version (2.1.302) as soon as possible but posting this question in the meantime. Does anyone know what's going on here? Thanks
.net-core, rhel, exit-code
3
1,665
1
https://stackoverflow.com/questions/51818610/net-core-application-exit-code-different-behaviour-on-linux-rhel-7
29,155,407
Connection to pypi.python.org timed out
i am managing a rhel server 6.x which have the Python-2.7.8 installed for an application. When the below chef resource is executed by the chef-client. execute "/usr/local/bin/python2.7 #{node[:base_dir]}/get-pip.py" I see the below error in the chef-client log. [2015-02-24T17:42:27+00:00] ERROR: execute[/usr/local/bin/python2.7 /opt/mount/get-pip.py] (python_app::default line 105) had an error: Mixlib::ShellOut::ShellCommandFailed: Expected process to exit with [0], but received '1' ---- Begin output of /usr/local/bin/python2.7 /opt/mount1/get-pip.py ---- STDOUT: Retrying (Retry(total=4, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee268d0>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=3, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee269d0>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=2, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee26ad0>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=1, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee26bd0>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=0, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee26cd0>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=4, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee26f50>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=3, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152efc4150>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=2, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152efc4250>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=1, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152efc4350>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=0, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152efc4450>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Could not find any downloads that satisfy the requirement pip in /usr/local/lib/python2.7/site-packages Collecting pip No distributions at all found for pip in /usr/local/lib/python2.7/site-packages STDERR: ---- End output of /usr/local/bin/python2.7 /opt/mount1/get-pip.py ---- I am looking up python for this issue. I noticed that there is another version of python. i think this comes with the rhel install. Python 2.6.6 The python 2.7.8 was install using chef as below: ark 'python' do url '[URL] path "/opt/mount/" owner 'python' action :put end execute "cd #{node[:base_dir]}/python && ./configure && make && make altinstall" do not_if { File.exists?('/usr/local/bin/python2.7') } end One thing that i observed is the '/opt/mount/python' is owned by 'root' as against 'python' user. Any idea why this error messages?
Connection to pypi.python.org timed out i am managing a rhel server 6.x which have the Python-2.7.8 installed for an application. When the below chef resource is executed by the chef-client. execute "/usr/local/bin/python2.7 #{node[:base_dir]}/get-pip.py" I see the below error in the chef-client log. [2015-02-24T17:42:27+00:00] ERROR: execute[/usr/local/bin/python2.7 /opt/mount/get-pip.py] (python_app::default line 105) had an error: Mixlib::ShellOut::ShellCommandFailed: Expected process to exit with [0], but received '1' ---- Begin output of /usr/local/bin/python2.7 /opt/mount1/get-pip.py ---- STDOUT: Retrying (Retry(total=4, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee268d0>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=3, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee269d0>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=2, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee26ad0>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=1, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee26bd0>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=0, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee26cd0>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=4, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152ee26f50>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=3, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152efc4150>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=2, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152efc4250>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=1, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152efc4350>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Retrying (Retry(total=0, connect=None, read=None, redirect=None)) after connection broken by 'ConnectTimeoutError(<pip._vendor.requests.packages.urllib3.connection.VerifiedHTTPSConnection object at 0x7f152efc4450>, 'Connection to pypi.python.org timed out. (connect timeout=15)')': /simple/pip/ Could not find any downloads that satisfy the requirement pip in /usr/local/lib/python2.7/site-packages Collecting pip No distributions at all found for pip in /usr/local/lib/python2.7/site-packages STDERR: ---- End output of /usr/local/bin/python2.7 /opt/mount1/get-pip.py ---- I am looking up python for this issue. I noticed that there is another version of python. i think this comes with the rhel install. Python 2.6.6 The python 2.7.8 was install using chef as below: ark 'python' do url '[URL] path "/opt/mount/" owner 'python' action :put end execute "cd #{node[:base_dir]}/python && ./configure && make && make altinstall" do not_if { File.exists?('/usr/local/bin/python2.7') } end One thing that i observed is the '/opt/mount/python' is owned by 'root' as against 'python' user. Any idea why this error messages?
python, chef-infra, rhel
3
25,341
1
https://stackoverflow.com/questions/29155407/connection-to-pypi-python-org-timed-out
8,683,724
Slow initial timing results using gettimeofday - worse under RHEL6 Server
I am using gettimeofday() to time a simple matrix multiply example, but I'm getting results that are close to twice too long initially. On a RHEL6 Server machine, I'm getting "bad" timing results for up to nearly 1 second (~65 individual timings in this example). All our other machines are RHEL5 Workstation boxes and this code works much better on them; I only get a couple of "bad" results initially (for first ~20 milliseconds). From posts on this site, I think this probably has something to do with the OS process scheduler. If I uncomment the first "for" statement below (thereby inserting an initial busy loop by repeatedly initializing the matrices a, b and c), I get zero "bad" results under both RHEL5 Workstation and RHEL6 Server. Alternatively, if I uncomment the sleep statement, I get ALL "bad" timing results for both RHEL5 & RHEL6. For some reason, my process is starting up with only about half the access to the CPU initially, then it gets "full" access to the CPU as long as the process stays busy. If it "sleeps" and then resumes timing, it again is temporarily only getting about half the full access to the CPU. Nothing else is happening on the machine (X is not running). I have tried "chrt" to control the priority of the process, but that changed nothing. I've verified this occurs with both GCC 4.4.6 and ICC 12.1.0. I've tried "nice" as well. Here's the code: #include <stdio.h> #include <unistd.h> #include <sys/time.h> #define N 225 #define DELAY_LOOPS 8000 main() { struct timeval _t0, _t1, _t2; double a[N][N], b[N][N], c[N][N]; double millisec, cum_ms; int i, j, k, l, m=0; gettimeofday( &_t0, NULL ); // for( l=0; l<DELAY_LOOPS; l++ ) for( i=0; i<N; i++ ) for( j=0; j<N; j++ ) { a[i][j]=0; b[i][j]=i; c[i][j]=j; } for( l=0; l<75; l++ ) { gettimeofday( &_t1, NULL ); for( i=0; i<N; i++ ) for( j=0; j<N; j++ ) for( k=0; k<N; k++ ) a[i][j]+=b[i][k]*c[k][j]; gettimeofday( &_t2, NULL ); millisec = 1000*(_t2.tv_sec-_t1.tv_sec); millisec += 1e-3*(_t2.tv_usec-_t1.tv_usec); cum_ms = 1000*(_t2.tv_sec-_t0.tv_sec); cum_ms += 1e-3*(_t2.tv_usec-_t0.tv_usec); printf( "%d: duration %fms, cumulative %fms\n", m++, millisec, cum_ms ); // sleep( 2 ); } printf( "a[%d][%d]=%f\n", N/2, N/2, a[N/2][N/2] ); } and here are the results: % icc -O2 -o test main.c; ./test 0: duration 13.049000ms, cumulative 13.677000ms 1: duration 13.026000ms, cumulative 26.753000ms 2: duration 12.911000ms, cumulative 39.668000ms 3: duration 12.913000ms, cumulative 52.584000ms 4: duration 12.914000ms, cumulative 65.501000ms 5: duration 12.911000ms, cumulative 78.415000ms 6: duration 12.912000ms, cumulative 91.331000ms /* snip */ 64: duration 12.912000ms, cumulative 840.633000ms 65: duration 10.455000ms, cumulative 851.092000ms 66: duration 5.910000ms, cumulative 857.004000ms 67: duration 5.908000ms, cumulative 862.914000ms 68: duration 5.907000ms, cumulative 868.823000ms 69: duration 5.908000ms, cumulative 874.732000ms 70: duration 5.912000ms, cumulative 880.646000ms 71: duration 5.907000ms, cumulative 886.554000ms 72: duration 5.907000ms, cumulative 892.462000ms 73: duration 5.908000ms, cumulative 898.372000ms 74: duration 5.908000ms, cumulative 904.281000ms a[112][112]=211680000.000000 I experience the problem regardless of optimization level (-O0, -O1, -O2, etc.). Does anyone know anything about how scheduling is done under RHEL6 Server? Is it so very different from RHEL5 Workstation? I presume the difference I'm seeing is more a function of the fact that one box is a server edition of RHEL and the other is a workstation edition (rather than the difference between release 5 vs. 6). Is there some simple way to reduce this effect under RHEL6 Server and make it act more like the RHEL5 Workstation boxes? Any ideas? Thanks.
Slow initial timing results using gettimeofday - worse under RHEL6 Server I am using gettimeofday() to time a simple matrix multiply example, but I'm getting results that are close to twice too long initially. On a RHEL6 Server machine, I'm getting "bad" timing results for up to nearly 1 second (~65 individual timings in this example). All our other machines are RHEL5 Workstation boxes and this code works much better on them; I only get a couple of "bad" results initially (for first ~20 milliseconds). From posts on this site, I think this probably has something to do with the OS process scheduler. If I uncomment the first "for" statement below (thereby inserting an initial busy loop by repeatedly initializing the matrices a, b and c), I get zero "bad" results under both RHEL5 Workstation and RHEL6 Server. Alternatively, if I uncomment the sleep statement, I get ALL "bad" timing results for both RHEL5 & RHEL6. For some reason, my process is starting up with only about half the access to the CPU initially, then it gets "full" access to the CPU as long as the process stays busy. If it "sleeps" and then resumes timing, it again is temporarily only getting about half the full access to the CPU. Nothing else is happening on the machine (X is not running). I have tried "chrt" to control the priority of the process, but that changed nothing. I've verified this occurs with both GCC 4.4.6 and ICC 12.1.0. I've tried "nice" as well. Here's the code: #include <stdio.h> #include <unistd.h> #include <sys/time.h> #define N 225 #define DELAY_LOOPS 8000 main() { struct timeval _t0, _t1, _t2; double a[N][N], b[N][N], c[N][N]; double millisec, cum_ms; int i, j, k, l, m=0; gettimeofday( &_t0, NULL ); // for( l=0; l<DELAY_LOOPS; l++ ) for( i=0; i<N; i++ ) for( j=0; j<N; j++ ) { a[i][j]=0; b[i][j]=i; c[i][j]=j; } for( l=0; l<75; l++ ) { gettimeofday( &_t1, NULL ); for( i=0; i<N; i++ ) for( j=0; j<N; j++ ) for( k=0; k<N; k++ ) a[i][j]+=b[i][k]*c[k][j]; gettimeofday( &_t2, NULL ); millisec = 1000*(_t2.tv_sec-_t1.tv_sec); millisec += 1e-3*(_t2.tv_usec-_t1.tv_usec); cum_ms = 1000*(_t2.tv_sec-_t0.tv_sec); cum_ms += 1e-3*(_t2.tv_usec-_t0.tv_usec); printf( "%d: duration %fms, cumulative %fms\n", m++, millisec, cum_ms ); // sleep( 2 ); } printf( "a[%d][%d]=%f\n", N/2, N/2, a[N/2][N/2] ); } and here are the results: % icc -O2 -o test main.c; ./test 0: duration 13.049000ms, cumulative 13.677000ms 1: duration 13.026000ms, cumulative 26.753000ms 2: duration 12.911000ms, cumulative 39.668000ms 3: duration 12.913000ms, cumulative 52.584000ms 4: duration 12.914000ms, cumulative 65.501000ms 5: duration 12.911000ms, cumulative 78.415000ms 6: duration 12.912000ms, cumulative 91.331000ms /* snip */ 64: duration 12.912000ms, cumulative 840.633000ms 65: duration 10.455000ms, cumulative 851.092000ms 66: duration 5.910000ms, cumulative 857.004000ms 67: duration 5.908000ms, cumulative 862.914000ms 68: duration 5.907000ms, cumulative 868.823000ms 69: duration 5.908000ms, cumulative 874.732000ms 70: duration 5.912000ms, cumulative 880.646000ms 71: duration 5.907000ms, cumulative 886.554000ms 72: duration 5.907000ms, cumulative 892.462000ms 73: duration 5.908000ms, cumulative 898.372000ms 74: duration 5.908000ms, cumulative 904.281000ms a[112][112]=211680000.000000 I experience the problem regardless of optimization level (-O0, -O1, -O2, etc.). Does anyone know anything about how scheduling is done under RHEL6 Server? Is it so very different from RHEL5 Workstation? I presume the difference I'm seeing is more a function of the fact that one box is a server edition of RHEL and the other is a workstation edition (rather than the difference between release 5 vs. 6). Is there some simple way to reduce this effect under RHEL6 Server and make it act more like the RHEL5 Workstation boxes? Any ideas? Thanks.
c, linux, scheduler, rhel, gettimeofday
3
921
2
https://stackoverflow.com/questions/8683724/slow-initial-timing-results-using-gettimeofday-worse-under-rhel6-server
7,066,761
.jar works in Windows but not Linux (stacktrace)
This is what I get: [user@localhost KioskMenu]$ java -jar ./kioskmenu.jar Exception in thread "main" java.lang.ClassFormatError: KioskMenu (unrecognized class file version) at java.lang.VMClassLoader.defineClass(libgcj.so.7rh) at java.lang.ClassLoader.defineClass(libgcj.so.7rh) at java.security.SecureClassLoader.defineClass(libgcj.so.7rh) at java.net.URLClassLoader.findClass(libgcj.so.7rh) at gnu.gcj.runtime.SystemClassLoader.findClass(libgcj.so.7rh) at java.lang.ClassLoader.loadClass(libgcj.so.7rh) at java.lang.ClassLoader.loadClass(libgcj.so.7rh) at gnu.java.lang.MainThread.run(libgcj.so.7rh) [user@localhost KioskMenu]$ The only thing I noticed was that my RHEL is using a "free" version of Java. But this shouldn't matter should it?
.jar works in Windows but not Linux (stacktrace) This is what I get: [user@localhost KioskMenu]$ java -jar ./kioskmenu.jar Exception in thread "main" java.lang.ClassFormatError: KioskMenu (unrecognized class file version) at java.lang.VMClassLoader.defineClass(libgcj.so.7rh) at java.lang.ClassLoader.defineClass(libgcj.so.7rh) at java.security.SecureClassLoader.defineClass(libgcj.so.7rh) at java.net.URLClassLoader.findClass(libgcj.so.7rh) at gnu.gcj.runtime.SystemClassLoader.findClass(libgcj.so.7rh) at java.lang.ClassLoader.loadClass(libgcj.so.7rh) at java.lang.ClassLoader.loadClass(libgcj.so.7rh) at gnu.java.lang.MainThread.run(libgcj.so.7rh) [user@localhost KioskMenu]$ The only thing I noticed was that my RHEL is using a "free" version of Java. But this shouldn't matter should it?
java, linux, jar, rhel
3
2,273
2
https://stackoverflow.com/questions/7066761/jar-works-in-windows-but-not-linux-stacktrace
58,855,758
Cannot run docker commands through Jenkin&#39;s Blue Ocean: ERRO[0000] No subuid ranges found for user “jenkins” in /etc/subuid
Setup my first pipeline using Jenkin's Ocean Blue on a EC2 RHEL 8 instance and when attempting to run docker commands (such as as docker pull maven:3.6.2-jdk-14) as the jenkins user the following error is always returned No subuid ranges found for user "jenkins" in /etc/subuid From a bit of research it seems I need to add the jenkins user to /etc/subuid and /etc/subgid but I'm unsure how to do this or what to add. Any help will be much appreciated.
Cannot run docker commands through Jenkin&#39;s Blue Ocean: ERRO[0000] No subuid ranges found for user “jenkins” in /etc/subuid Setup my first pipeline using Jenkin's Ocean Blue on a EC2 RHEL 8 instance and when attempting to run docker commands (such as as docker pull maven:3.6.2-jdk-14) as the jenkins user the following error is always returned No subuid ranges found for user "jenkins" in /etc/subuid From a bit of research it seems I need to add the jenkins user to /etc/subuid and /etc/subgid but I'm unsure how to do this or what to add. Any help will be much appreciated.
docker, jenkins, rhel, jenkins-blueocean
3
6,473
1
https://stackoverflow.com/questions/58855758/cannot-run-docker-commands-through-jenkins-blue-ocean-erro0000-no-subuid-ran
15,890,481
Selinux - File Contexts Look Good, But Selinux Won&#39;t Allow Write
I am trying to learn Selinux. With a sandbox and using VSFTPD to experiment with, I have a vsfptd server running in Centos. I have annonmous users to place files in /var/ftp/incoming. On a remote machine I can have the user successfully log in but could not place the file on the remove vsftpd server: $ftp mysql_server Connected to mysql_server (192.168.1.31). 220 Welcome to blah FTP service. Name (mysql_server:root): anonymous 331 Please specify the password. Password: 230 Login successful. Remote system type is UNIX. Using binary mode to transfer ftp> put atd local: atd remote: atd 227 Entering Passive Mode (192,168,1,31,19,161). 553 Could not create file. ftp> On the VSFTPD server, aureport -a report shows: [root@mysql_server ftp]# aureport -a AVC Report ======================================================== # date time comm subj syscall class permission obj event ======================================================== 4. 04/08/2013 13:30:36 vsftpd unconfined_u:system_r:ftpd_t:s0-s0:c0.c1023 21 dir write system_u:object_r:public_content_t:s0 denied 28 5. 04/08/2013 13:34:57 vsftpd unconfined_u:system_r:ftpd_t:s0-s0:c0.c1023 2 dir write system_u:object_r:public_content_t:s0 denied 47 I checked the directory and the file contexts look good, so I don't understand why Selinux won't allow vsftpd to write to the incoming directory: [root@mysql_server ftp]# ls -Z drwx-wx---. root ftp system_u:object_r:public_content_t:s0 incoming drwxr-xr-x. root root system_u:object_r:public_content_t:s0 pub [root@mysql_server ftp]#
Selinux - File Contexts Look Good, But Selinux Won&#39;t Allow Write I am trying to learn Selinux. With a sandbox and using VSFTPD to experiment with, I have a vsfptd server running in Centos. I have annonmous users to place files in /var/ftp/incoming. On a remote machine I can have the user successfully log in but could not place the file on the remove vsftpd server: $ftp mysql_server Connected to mysql_server (192.168.1.31). 220 Welcome to blah FTP service. Name (mysql_server:root): anonymous 331 Please specify the password. Password: 230 Login successful. Remote system type is UNIX. Using binary mode to transfer ftp> put atd local: atd remote: atd 227 Entering Passive Mode (192,168,1,31,19,161). 553 Could not create file. ftp> On the VSFTPD server, aureport -a report shows: [root@mysql_server ftp]# aureport -a AVC Report ======================================================== # date time comm subj syscall class permission obj event ======================================================== 4. 04/08/2013 13:30:36 vsftpd unconfined_u:system_r:ftpd_t:s0-s0:c0.c1023 21 dir write system_u:object_r:public_content_t:s0 denied 28 5. 04/08/2013 13:34:57 vsftpd unconfined_u:system_r:ftpd_t:s0-s0:c0.c1023 2 dir write system_u:object_r:public_content_t:s0 denied 47 I checked the directory and the file contexts look good, so I don't understand why Selinux won't allow vsftpd to write to the incoming directory: [root@mysql_server ftp]# ls -Z drwx-wx---. root ftp system_u:object_r:public_content_t:s0 incoming drwxr-xr-x. root root system_u:object_r:public_content_t:s0 pub [root@mysql_server ftp]#
centos, rhel, selinux, security-context
3
3,143
2
https://stackoverflow.com/questions/15890481/selinux-file-contexts-look-good-but-selinux-wont-allow-write
9,626,960
Building FFmpeg on RHEL4
StackOverflow has yet to fail me, so don't let me down this time! I need to build FFmpeg on RHEL4 ... yes, 4! The version of make on RHEL4 is 3.80 , and the Makefile that ./configure generates when configuring FFmpeg does not seem to be compatible. When I make it throws a parse error every time. This happens on pretty much every software pkg I try and build. My applications group (at work) will not upgrade this machine, so please skip the suggestions to upgrade to RHEL5 or 6. What are my options for installing FFmpeg?
Building FFmpeg on RHEL4 StackOverflow has yet to fail me, so don't let me down this time! I need to build FFmpeg on RHEL4 ... yes, 4! The version of make on RHEL4 is 3.80 , and the Makefile that ./configure generates when configuring FFmpeg does not seem to be compatible. When I make it throws a parse error every time. This happens on pretty much every software pkg I try and build. My applications group (at work) will not upgrade this machine, so please skip the suggestions to upgrade to RHEL5 or 6. What are my options for installing FFmpeg?
ffmpeg, makefile, rhel
3
2,184
1
https://stackoverflow.com/questions/9626960/building-ffmpeg-on-rhel4
79,621,916
Malformed Block in Ansible
I'm trying to run a playbook that gathers yum history logs, and consolidates them on a single server. I ran this playbook --- - name: Gather yum history from all hosts hosts: all become: true tasks: - name: Run 'yum history' command shell: yum history register: yum_history_output - name: Save yum history to local file copy: content: | Host: {{ inventory_hostname }} ======================================== {{ yum_history_output.stdout }} ======================================== dest: "/tmp/yum_history_{{ inventory_hostname }}.log" - name: Fetch all yum history logs to control node hosts: all gather_facts: false tasks: - name: Fetch log files to control node fetch: src: "/tmp/yum_history_{{ inventory_hostname }}.log" dest: "./yum_history_logs/" flat: yes and, after some formatting, got a malformed block error which just seemed to put the whole playbook into a single block. What am I doing wrong? I was given a bunch of formatting issues that I fixed, mostly deleting a lot of spaces because it didn't like the spaces in vi, but now I get a malformed block, and have no idea what to do next.
Malformed Block in Ansible I'm trying to run a playbook that gathers yum history logs, and consolidates them on a single server. I ran this playbook --- - name: Gather yum history from all hosts hosts: all become: true tasks: - name: Run 'yum history' command shell: yum history register: yum_history_output - name: Save yum history to local file copy: content: | Host: {{ inventory_hostname }} ======================================== {{ yum_history_output.stdout }} ======================================== dest: "/tmp/yum_history_{{ inventory_hostname }}.log" - name: Fetch all yum history logs to control node hosts: all gather_facts: false tasks: - name: Fetch log files to control node fetch: src: "/tmp/yum_history_{{ inventory_hostname }}.log" dest: "./yum_history_logs/" flat: yes and, after some formatting, got a malformed block error which just seemed to put the whole playbook into a single block. What am I doing wrong? I was given a bunch of formatting issues that I fixed, mostly deleting a lot of spaces because it didn't like the spaces in vi, but now I get a malformed block, and have no idea what to do next.
automation, ansible, rhel
3
49
1
https://stackoverflow.com/questions/79621916/malformed-block-in-ansible
73,196,959
Trying to install Docker on RHEL - docker-ce conflicts with Podman-docker
Getting this error: Updating Subscription Management repositories. Red Hat Enterprise Linux 8 for x86_64 - AppStream (RPMs) 22 MB/s | 45 MB 00:02 Last metadata expiration check: 0:00:01 ago on Mon 01 Aug 2022 17:36:44 BST. Error: Problem: problem with installed package podman-docker-2:4.0.2-6.module+el8.6.0+14877+f643d2d6.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-2:4.0.2-6.module+el8.6.0+14877+f643d2d6.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.0.0-2.git921f98f.module+el8+2785+ff8a053f.noarch package podman-docker-1.0.0-2.git921f98f.module+el8+2785+ff8a053f.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.0.0-2.git921f98f.module+el8.0.0+2958+4e823551.noarch package podman-docker-1.0.0-2.git921f98f.module+el8.0.0+2958+4e823551.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.0.5-1.gitf604175.module+el8.0.0+4017+bbba319f.noarch package podman-docker-1.0.5-1.gitf604175.module+el8.0.0+4017+bbba319f.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.0.3-1.git9d78c0c.module+el8.0.0.z+3717+fdd07b7c.noarch package podman-docker-1.0.3-1.git9d78c0c.module+el8.0.0.z+3717+fdd07b7c.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.4.2-5.module+el8.1.0+4240+893c1ab8.noarch package podman-docker-1.4.2-5.module+el8.1.0+4240+893c1ab8.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.6.4-2.module+el8.1.1+5363+bf8ff1af.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.6.4-4.module+el8.1.1+5885+44006e55.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.6.4-10.module+el8.2.0+6063+e761893a.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.4.2-6.module+el8.1.0+4830+f49150d7.noarch package podman-docker-1.4.2-6.module+el8.1.0+4830+f49150d7.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.9.3-2.module+el8.2.1+6867+366c07d6.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.6.4-11.module+el8.2.0+6368+cf16aa14.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.6.4-12.module+el8.2.0+6669+dde598ec.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-2.0.5-5.module+el8.3.0+8221+97165c3f.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-2.2.1-7.module+el8.3.1+9857+68fb1526.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-3.0.1-6.module+el8.4.0+10607+f4da7515.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-3.0.1-7.module+el8.4.0+11311+9da8acfb.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-3.2.3-0.10.module+el8.4.0+11989+6676f7ad.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-3.2.3-0.11.module+el8.4.0+12050+ef972f71.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-3.3.1-9.module+el8.5.0+12697+018f24d7.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-2:4.0.2-6.module+el8.6.0+14673+621cb8be.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1:3.4.2-9.module+el8.5.0+13852+150547f7.noarch cannot install the best candidate for the job
Trying to install Docker on RHEL - docker-ce conflicts with Podman-docker Getting this error: Updating Subscription Management repositories. Red Hat Enterprise Linux 8 for x86_64 - AppStream (RPMs) 22 MB/s | 45 MB 00:02 Last metadata expiration check: 0:00:01 ago on Mon 01 Aug 2022 17:36:44 BST. Error: Problem: problem with installed package podman-docker-2:4.0.2-6.module+el8.6.0+14877+f643d2d6.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-2:4.0.2-6.module+el8.6.0+14877+f643d2d6.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.0.0-2.git921f98f.module+el8+2785+ff8a053f.noarch package podman-docker-1.0.0-2.git921f98f.module+el8+2785+ff8a053f.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.0.0-2.git921f98f.module+el8.0.0+2958+4e823551.noarch package podman-docker-1.0.0-2.git921f98f.module+el8.0.0+2958+4e823551.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.0.5-1.gitf604175.module+el8.0.0+4017+bbba319f.noarch package podman-docker-1.0.5-1.gitf604175.module+el8.0.0+4017+bbba319f.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.0.3-1.git9d78c0c.module+el8.0.0.z+3717+fdd07b7c.noarch package podman-docker-1.0.3-1.git9d78c0c.module+el8.0.0.z+3717+fdd07b7c.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.4.2-5.module+el8.1.0+4240+893c1ab8.noarch package podman-docker-1.4.2-5.module+el8.1.0+4240+893c1ab8.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.6.4-2.module+el8.1.1+5363+bf8ff1af.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.6.4-4.module+el8.1.1+5885+44006e55.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.6.4-10.module+el8.2.0+6063+e761893a.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1.4.2-6.module+el8.1.0+4830+f49150d7.noarch package podman-docker-1.4.2-6.module+el8.1.0+4830+f49150d7.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.9.3-2.module+el8.2.1+6867+366c07d6.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.6.4-11.module+el8.2.0+6368+cf16aa14.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package podman-docker-1.6.4-12.module+el8.2.0+6669+dde598ec.noarch conflicts with docker-ce provided by docker-ce-3:20.10.17-3.el8.x86_64 package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-2.0.5-5.module+el8.3.0+8221+97165c3f.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-2.2.1-7.module+el8.3.1+9857+68fb1526.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-3.0.1-6.module+el8.4.0+10607+f4da7515.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-3.0.1-7.module+el8.4.0+11311+9da8acfb.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-3.2.3-0.10.module+el8.4.0+11989+6676f7ad.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-3.2.3-0.11.module+el8.4.0+12050+ef972f71.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-3.3.1-9.module+el8.5.0+12697+018f24d7.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-2:4.0.2-6.module+el8.6.0+14673+621cb8be.noarch package docker-ce-3:20.10.17-3.el8.x86_64 conflicts with docker provided by podman-docker-1:3.4.2-9.module+el8.5.0+13852+150547f7.noarch cannot install the best candidate for the job
linux, docker, rhel, podman
3
4,293
1
https://stackoverflow.com/questions/73196959/trying-to-install-docker-on-rhel-docker-ce-conflicts-with-podman-docker
72,141,594
microdnf shows librhsm-WARNING Found 0 entitlement certificates
Problem I am trying to install nmap-ncat package into a container created with this image by microdnf pkg-manager. I run microdnf update microdnf -y install nmap-ncat Both cmds output (microdnf:xxx): librhsm-WARNING **: 12:00:00.000: Found 0 entitlement certificates ... but both cmds seem to be working properly after this. I can actually see ncat cmd's help by --help option so it seems the install is complete. I googled about this issue and found this . It seems since I am not a red hat subscriber, Red Hat Enterprise Linux(RHEL) which is the OS used by the image limits what I can do. Maybe the base image of the image is one of red hat universal base images (ubi) by which allow commercial and open source developers to build containers based on RHEL without requiring them or their users to be RHEL subscribers (source) because the output of microdnf -y install nmap-ncat is like Installing: ... ;ubi-8-baseos Installing: ... ;ubi-8-appstream and this shows both ubi-8-baseos and ubi-8-appstream are RHEL 8 UBI repositories (but I am not really sure about this). My questions are Why did the warning come up while I successfully installed nmap-ncat? Am I entitled to use nmap-ncat with this image ?
microdnf shows librhsm-WARNING Found 0 entitlement certificates Problem I am trying to install nmap-ncat package into a container created with this image by microdnf pkg-manager. I run microdnf update microdnf -y install nmap-ncat Both cmds output (microdnf:xxx): librhsm-WARNING **: 12:00:00.000: Found 0 entitlement certificates ... but both cmds seem to be working properly after this. I can actually see ncat cmd's help by --help option so it seems the install is complete. I googled about this issue and found this . It seems since I am not a red hat subscriber, Red Hat Enterprise Linux(RHEL) which is the OS used by the image limits what I can do. Maybe the base image of the image is one of red hat universal base images (ubi) by which allow commercial and open source developers to build containers based on RHEL without requiring them or their users to be RHEL subscribers (source) because the output of microdnf -y install nmap-ncat is like Installing: ... ;ubi-8-baseos Installing: ... ;ubi-8-appstream and this shows both ubi-8-baseos and ubi-8-appstream are RHEL 8 UBI repositories (but I am not really sure about this). My questions are Why did the warning come up while I successfully installed nmap-ncat? Am I entitled to use nmap-ncat with this image ?
docker, rhel, docker-image
3
7,916
1
https://stackoverflow.com/questions/72141594/microdnf-shows-librhsm-warning-found-0-entitlement-certificates
62,373,716
Retrieve executable from core dump
I want to retrieve the executable from a core dump and the output of any linux package used to get this information should contain execfn in it's output. Here are the following things which I have tried so far : $ file kms kms: ELF 64-bit LSB core file x86-64, version 1 (SYSV), SVR4-style, from '/test', real uid: 1000440000, effective uid: 1000440000, real gid: 0, effective gid: 0, execfn: '/test', platform: 'x86_64' The file command only works for specific cores and it's not a generic solution because some core dump gives following output. $ file ss ss: ELF 64-bit LSB core file x86-64, version 1 (SYSV), too many program header sections (6841) gdb command doesn't work for all core dumps in the same manner. The output using gdb command is inconsistent. The output received by gdb command for some core dump is not the same as strings command. $gdb kms Core was generated by `/test'. I even tried strings package and I think it gives proper output but the format doesn't contain execfn for it to be used in my solution $ strings kms | grep ^/ | tail -1 /test Can anyone please suggest any linux package which will help me in retrieving executable from core dump which contains execfn in it's output.
Retrieve executable from core dump I want to retrieve the executable from a core dump and the output of any linux package used to get this information should contain execfn in it's output. Here are the following things which I have tried so far : $ file kms kms: ELF 64-bit LSB core file x86-64, version 1 (SYSV), SVR4-style, from '/test', real uid: 1000440000, effective uid: 1000440000, real gid: 0, effective gid: 0, execfn: '/test', platform: 'x86_64' The file command only works for specific cores and it's not a generic solution because some core dump gives following output. $ file ss ss: ELF 64-bit LSB core file x86-64, version 1 (SYSV), too many program header sections (6841) gdb command doesn't work for all core dumps in the same manner. The output using gdb command is inconsistent. The output received by gdb command for some core dump is not the same as strings command. $gdb kms Core was generated by `/test'. I even tried strings package and I think it gives proper output but the format doesn't contain execfn for it to be used in my solution $ strings kms | grep ^/ | tail -1 /test Can anyone please suggest any linux package which will help me in retrieving executable from core dump which contains execfn in it's output.
linux, executable, rhel, coredump
3
4,025
2
https://stackoverflow.com/questions/62373716/retrieve-executable-from-core-dump
54,296,427
Systemd exits with status=209/STDOUT
Scenario I have a systemd file which I want to run conditionally, but only if the environment variable ISCAPTUREPOD is set to true. I have a container which runs two services on startup, but there is a special scenario where I only want one of the services to run. I'm passing an environment variable via Kubernetes which I want to use to control whether the second service does or does not start. I have a script named iscapturepod.sh which checks the environment variable and it is part of an ExecStartPre statement. I want the script to succeed when the environment variable ISCAPTUREPOD is set to "True" and fail if it doesn't exist or is set to something other than "True". Problem: No matter what I do ExecStartPre fails. I have even tried just making the script say exit 0 . That's the only thing in the entire script just because I wanted to force a success. Systemd still fails with status 209/STDOUT. Moloch capture service: [Unit] Description=Moloch Capture After=network.target [Service] Type=simple Restart=on-failure StandardOutput=tty ExecStartPre= /bin/sh -c '/data/moloch/bin/iscapturepod.sh' ExecStart=/bin/sh -c '/data/moloch/bin/moloch-capture -c MOLOCH_INSTALL_DIR/etc/config.ini ${OPTIONS} >> /data/moloch/logs/capture.log 2>&1' LimitCORE=infinity LimitMEMLOCK=infinity [Install] WantedBy=multi-user.target The Script #!/bin/bash # This script checks to see whether this pod is or is not a capture pod # Kubernetes will pass the ISCAPTUREPOD variable as an environment variable with # value True to those pods meant for capture and False for the viewer pod. # This allows us to only use one container for both the viewer and capture pods # The molochcapture service will run this in an ExecStartPre statement. If it # throws an error this will prevent the molochcapture service from starting if [[ ! -z "${ISCAPTUREPOD}" ]]; then if [[ "${ISCAPTUREPOD}" == "True" ]]; then echo This is a capture pod exit 0 else echo This is not a capture pod 1>&2 exit 1 fi else echo This is not a capture pod 1>$2 exit 1 fi According to this site 0 should be a success. However, even if I change the script to just be exit 0 I still get: [root@sensor1 /]# systemctl status molochcapture ● molochcapture.service - Moloch Capture Loaded: loaded (/usr/lib/systemd/system/molochcapture.service; enabled; vendor preset: disabled) Active: failed (Result: start-limit) since Mon 2019-01-21 12:58:20 UTC; 1s ago Process: 281 ExecStartPre=/bin/sh -c /data/moloch/bin/iscapturepod.sh (code=exited, status=209/STDOUT) Jan 21 12:58:20 sensor1.lan systemd[1]: Failed to start Moloch Capture. Jan 21 12:58:20 sensor1.lan systemd[1]: Unit molochcapture.service entered failed state. Jan 21 12:58:20 sensor1.lan systemd[1]: molochcapture.service failed. Jan 21 12:58:20 sensor1.lan systemd[1]: molochcapture.service holdoff time over, scheduling restart. Jan 21 12:58:20 sensor1.lan systemd[1]: Stopped Moloch Capture. Jan 21 12:58:20 sensor1.lan systemd[1]: start request repeated too quickly for molochcapture.service Jan 21 12:58:20 sensor1.lan systemd[1]: Failed to start Moloch Capture. Jan 21 12:58:20 sensor1.lan systemd[1]: Unit molochcapture.service entered failed state. Jan 21 12:58:20 sensor1.lan systemd[1]: molochcapture.service failed. I have manually checked the script works correctly and has no problems. Kubernetes is passing the environment variable as expected and the script returns "This is a capture pod". I thought maybe it has something to do with systemd not having access to STDOUT, but that's when I tried just exit 0 and it still failed.
Systemd exits with status=209/STDOUT Scenario I have a systemd file which I want to run conditionally, but only if the environment variable ISCAPTUREPOD is set to true. I have a container which runs two services on startup, but there is a special scenario where I only want one of the services to run. I'm passing an environment variable via Kubernetes which I want to use to control whether the second service does or does not start. I have a script named iscapturepod.sh which checks the environment variable and it is part of an ExecStartPre statement. I want the script to succeed when the environment variable ISCAPTUREPOD is set to "True" and fail if it doesn't exist or is set to something other than "True". Problem: No matter what I do ExecStartPre fails. I have even tried just making the script say exit 0 . That's the only thing in the entire script just because I wanted to force a success. Systemd still fails with status 209/STDOUT. Moloch capture service: [Unit] Description=Moloch Capture After=network.target [Service] Type=simple Restart=on-failure StandardOutput=tty ExecStartPre= /bin/sh -c '/data/moloch/bin/iscapturepod.sh' ExecStart=/bin/sh -c '/data/moloch/bin/moloch-capture -c MOLOCH_INSTALL_DIR/etc/config.ini ${OPTIONS} >> /data/moloch/logs/capture.log 2>&1' LimitCORE=infinity LimitMEMLOCK=infinity [Install] WantedBy=multi-user.target The Script #!/bin/bash # This script checks to see whether this pod is or is not a capture pod # Kubernetes will pass the ISCAPTUREPOD variable as an environment variable with # value True to those pods meant for capture and False for the viewer pod. # This allows us to only use one container for both the viewer and capture pods # The molochcapture service will run this in an ExecStartPre statement. If it # throws an error this will prevent the molochcapture service from starting if [[ ! -z "${ISCAPTUREPOD}" ]]; then if [[ "${ISCAPTUREPOD}" == "True" ]]; then echo This is a capture pod exit 0 else echo This is not a capture pod 1>&2 exit 1 fi else echo This is not a capture pod 1>$2 exit 1 fi According to this site 0 should be a success. However, even if I change the script to just be exit 0 I still get: [root@sensor1 /]# systemctl status molochcapture ● molochcapture.service - Moloch Capture Loaded: loaded (/usr/lib/systemd/system/molochcapture.service; enabled; vendor preset: disabled) Active: failed (Result: start-limit) since Mon 2019-01-21 12:58:20 UTC; 1s ago Process: 281 ExecStartPre=/bin/sh -c /data/moloch/bin/iscapturepod.sh (code=exited, status=209/STDOUT) Jan 21 12:58:20 sensor1.lan systemd[1]: Failed to start Moloch Capture. Jan 21 12:58:20 sensor1.lan systemd[1]: Unit molochcapture.service entered failed state. Jan 21 12:58:20 sensor1.lan systemd[1]: molochcapture.service failed. Jan 21 12:58:20 sensor1.lan systemd[1]: molochcapture.service holdoff time over, scheduling restart. Jan 21 12:58:20 sensor1.lan systemd[1]: Stopped Moloch Capture. Jan 21 12:58:20 sensor1.lan systemd[1]: start request repeated too quickly for molochcapture.service Jan 21 12:58:20 sensor1.lan systemd[1]: Failed to start Moloch Capture. Jan 21 12:58:20 sensor1.lan systemd[1]: Unit molochcapture.service entered failed state. Jan 21 12:58:20 sensor1.lan systemd[1]: molochcapture.service failed. I have manually checked the script works correctly and has no problems. Kubernetes is passing the environment variable as expected and the script returns "This is a capture pod". I thought maybe it has something to do with systemd not having access to STDOUT, but that's when I tried just exit 0 and it still failed.
systemd, rhel
3
7,611
2
https://stackoverflow.com/questions/54296427/systemd-exits-with-status-209-stdout
36,760,111
How do i specify a requirement for a minimum version of openssl in rpm spec file?
The rpm spec file for my package looks like this: Requires:mysql-server >= 5.5.30, java >= 1.6, p7zip, openssl >= 1.0.2g I have an RHEL system where the OpenSSL version is this: [ag@rhel1]# openssl OpenSSL> version OpenSSL 1.0.1e-fips 11 Feb 2013 [ag@rhel1]# rpm -q --whatprovides openssl openssl-1.0.1e-51.el7_2.4.x86_64 When I try to test the install of my rpm package, it is supposed to fail because my package requires a minimum of 1.0.2g whereas 1.0.1e is installed. However, the installation succeeds without any problem. What am I missing here?
How do i specify a requirement for a minimum version of openssl in rpm spec file? The rpm spec file for my package looks like this: Requires:mysql-server >= 5.5.30, java >= 1.6, p7zip, openssl >= 1.0.2g I have an RHEL system where the OpenSSL version is this: [ag@rhel1]# openssl OpenSSL> version OpenSSL 1.0.1e-fips 11 Feb 2013 [ag@rhel1]# rpm -q --whatprovides openssl openssl-1.0.1e-51.el7_2.4.x86_64 When I try to test the install of my rpm package, it is supposed to fail because my package requires a minimum of 1.0.2g whereas 1.0.1e is installed. However, the installation succeeds without any problem. What am I missing here?
linux, openssl, rpm, rhel, rpm-spec
3
1,417
2
https://stackoverflow.com/questions/36760111/how-do-i-specify-a-requirement-for-a-minimum-version-of-openssl-in-rpm-spec-file
32,722,241
add rule to firewalld in Centos7 to allow all traffic from a server
I have a small cluster with Centos7. I'm trying how to use the new firewalld. I need a rule to allow all traffic between those servers. I was able to do it with: sudo iptables -A INPUT -s [hostname] -j ACCEPT and it worked. But now I have to use firewall-cmd because of Centos 7. How can I add a rule to allow all traffic between my nodes? I'm trying to run MPI on them but the firewalld is rejecting the connection so the solution I thought of came to this. My current firewall-cmd configuration is: $ firewall-cmd --list-all work (default, active) interfaces: eno1 sources: services: dhcpv6-client ipp-client ssh ports: masquerade: no forward-ports: icmp-blocks: rich rules:
add rule to firewalld in Centos7 to allow all traffic from a server I have a small cluster with Centos7. I'm trying how to use the new firewalld. I need a rule to allow all traffic between those servers. I was able to do it with: sudo iptables -A INPUT -s [hostname] -j ACCEPT and it worked. But now I have to use firewall-cmd because of Centos 7. How can I add a rule to allow all traffic between my nodes? I'm trying to run MPI on them but the firewalld is rejecting the connection so the solution I thought of came to this. My current firewall-cmd configuration is: $ firewall-cmd --list-all work (default, active) interfaces: eno1 sources: services: dhcpv6-client ipp-client ssh ports: masquerade: no forward-ports: icmp-blocks: rich rules:
firewall, iptables, rhel, centos7
3
21,553
2
https://stackoverflow.com/questions/32722241/add-rule-to-firewalld-in-centos7-to-allow-all-traffic-from-a-server
27,014,365
Git installation RHEL Linux failing with &quot;warning: zlib.h: No such file or directory&quot;
I am installing git following the document. But I am getting below error after running make command [URL] [root@liaan17 git-1.8.2.3]# make CC credential-store.o In file included from credential-store.c:1: cache.h:19:18: warning: zlib.h: No such file or directory In file included from credential-store.c:1: cache.h:21: error: expected specifier-qualifier-list before ‘z_stream’
Git installation RHEL Linux failing with &quot;warning: zlib.h: No such file or directory&quot; I am installing git following the document. But I am getting below error after running make command [URL] [root@liaan17 git-1.8.2.3]# make CC credential-store.o In file included from credential-store.c:1: cache.h:19:18: warning: zlib.h: No such file or directory In file included from credential-store.c:1: cache.h:21: error: expected specifier-qualifier-list before ‘z_stream’
git, zlib, rhel, failed-installation
3
8,443
2
https://stackoverflow.com/questions/27014365/git-installation-rhel-linux-failing-with-warning-zlib-h-no-such-file-or-direc
61,403,052
What causes this error message when using the postgres pg_dump command?
I am attempting to dump an entire PG Database from hosting on a GCP/GCE instance ( RHEL 8.x VM ). The purpose is to transfer the .sql file to another RHEL 8.x VM on GCE and restore the database. I am running this script as OS user 'postgres' #!/bin/bash clear export PGUSER=<pg_user> export PG_DATABASE=<pg_db> export PGHOSTADDR=<pg_ip_addr> export PGPORT=5432 SQL_FILE=<pg_db>.sql export PGPASSWORD=<passwd> pg_dump \ --format plain \ --no-owner \ --verbose \ --file $SQL_FILE \ $PG_DATABASE If there is anything else recommended in a script to dump a database to a .sql file for restoration: anything recommended adding to the script, or any way to improve the script that you are aware of), please advise. I get this error when running the script: (have PostGIS installed) pg_dump: last built-in OID is 16383 pg_dump: reading extensions pg_dump: identifying extension members pg_dump: reading schemas pg_dump: reading user-defined tables pg_dump: error: query failed: ERROR: permission denied for table us_lex pg_dump: error: query was: LOCK TABLE public.us_lex IN ACCESS SHARE MODE Why is permission denied when running as OS user postgres . What permission would I need to add (full/exact command), to add whatever permissions are needed? Or how could I exclude this table? I don't think I need it, and it would be recreated with the installation of a new Postgres instance.
What causes this error message when using the postgres pg_dump command? I am attempting to dump an entire PG Database from hosting on a GCP/GCE instance ( RHEL 8.x VM ). The purpose is to transfer the .sql file to another RHEL 8.x VM on GCE and restore the database. I am running this script as OS user 'postgres' #!/bin/bash clear export PGUSER=<pg_user> export PG_DATABASE=<pg_db> export PGHOSTADDR=<pg_ip_addr> export PGPORT=5432 SQL_FILE=<pg_db>.sql export PGPASSWORD=<passwd> pg_dump \ --format plain \ --no-owner \ --verbose \ --file $SQL_FILE \ $PG_DATABASE If there is anything else recommended in a script to dump a database to a .sql file for restoration: anything recommended adding to the script, or any way to improve the script that you are aware of), please advise. I get this error when running the script: (have PostGIS installed) pg_dump: last built-in OID is 16383 pg_dump: reading extensions pg_dump: identifying extension members pg_dump: reading schemas pg_dump: reading user-defined tables pg_dump: error: query failed: ERROR: permission denied for table us_lex pg_dump: error: query was: LOCK TABLE public.us_lex IN ACCESS SHARE MODE Why is permission denied when running as OS user postgres . What permission would I need to add (full/exact command), to add whatever permissions are needed? Or how could I exclude this table? I don't think I need it, and it would be recreated with the installation of a new Postgres instance.
postgresql, google-compute-engine, rhel, pg-dump
3
3,018
1
https://stackoverflow.com/questions/61403052/what-causes-this-error-message-when-using-the-postgres-pg-dump-command
59,653,487
How to install Python 3.8.1 on RHEL 8 UBI container with dnf?
How do I install and run Python 3.8.1 on Linux? Python.org does not offer a Linux binary for Python 3.8.1. I want to type: dnf -y install python or dnf -y install python38 and have it install Python 3.8.1 on my RHEL 8.0 or 8.1 inside my minimal container that I want to keep small. How would I make this happen? I don't even see a download for Python for Linux on python.org. Do people use Python on Linux? There are no Linux binaries to be downloaded from Python.org I have searched all over (internet search), and cannot find a way to install Python 3.8.1 easily on a minimal container. I do not want to download the source and compile, because it is a small container that I am trying to keep small without a lot of overhead. Reference: [URL]
How to install Python 3.8.1 on RHEL 8 UBI container with dnf? How do I install and run Python 3.8.1 on Linux? Python.org does not offer a Linux binary for Python 3.8.1. I want to type: dnf -y install python or dnf -y install python38 and have it install Python 3.8.1 on my RHEL 8.0 or 8.1 inside my minimal container that I want to keep small. How would I make this happen? I don't even see a download for Python for Linux on python.org. Do people use Python on Linux? There are no Linux binaries to be downloaded from Python.org I have searched all over (internet search), and cannot find a way to install Python 3.8.1 easily on a minimal container. I do not want to download the source and compile, because it is a small container that I am trying to keep small without a lot of overhead. Reference: [URL]
python, fedora, rhel
3
16,754
1
https://stackoverflow.com/questions/59653487/how-to-install-python-3-8-1-on-rhel-8-ubi-container-with-dnf
59,574,914
How to run podman commands on host from within container
In case of docker, this can be achieved by mounting docker.sock inside container. But since there is no daemon in podman. What's the replacement for docker.sock? I want to typically check the podman images presents on host and start a new container. I'm using Podman with --privileged=true and root.
How to run podman commands on host from within container In case of docker, this can be achieved by mounting docker.sock inside container. But since there is no daemon in podman. What's the replacement for docker.sock? I want to typically check the podman images presents on host and start a new container. I'm using Podman with --privileged=true and root.
docker, containers, rhel, podman
3
4,087
1
https://stackoverflow.com/questions/59574914/how-to-run-podman-commands-on-host-from-within-container
51,000,791
How to mount HostPath Volume in Kubernetes with SELinux
I am trying to mount a hostPath volume into a Kubernetes Pod. An example of a hostPath volume specification is shown below, which is taken from the docs. I am deploying to hosts that are running RHEL 7 with SELinux enabled. apiVersion: v1 kind: Pod metadata: name: test-pd spec: containers: - image: k8s.gcr.io/test-webserver name: test-container volumeMounts: - mountPath: /test-pd name: test-volume volumes: - name: test-volume hostPath: # directory location on host path: /data # this field is optional type: Directory When my Pod tries to read from a file that has been mounted from the underlying host, I get a "Permission Denied" error. When I run setenforce 0 to turn off SELinux, the error goes away and I can access the file. I get the same error when I bind mount a directory into a Docker container. The issue is described here and, when using Docker, can be fixed by using the z or Z bind mount flag, described in the Docker docs here . Whilst I can fix the issue by running chcon -Rt svirt_sandbox_file_t /path/to/my/host/dir/to/mount I see this as a nasty hack, as I need to do this on every host in my Kubernetes cluster and also because my deployment of Kubernetes as described in the YAML spec is not a complete description of what it is that needs to be done to get my YAML to run correctly. Turning off SELinux is not an option. I can see that Kubernetes mentions SELinux security contexts in the docs here , but I haven't been able to successfully mount a hostPath volume into a pod without getting the permission denied error. What does the YAML need to look like to successfully enable a container to mount a HostPath volume from an underlying host that is running SELinux? Update: The file I am accessing is a CA certificate that has these labels: system_u:object_r:cert_t:s0 When I use the following options: securityContext: seLinuxOptions: level: "s0:c123,c456" and then check the access control audit errors via ausearch -m avc -ts recent , I can see that there is a permission denied error where the container has a level label of s0:c123,c456 , so I can see that the level label works. I have set the label to be s0 . However, if I try to change the type label to be cert_t , the container doesn't even start, there's an error : container_linux.go:247: starting container process caused "process_linux.go:364: container init caused \"write /proc/self/task/1/attr/exec: invalid argument\"" I don't seem to be able to change the type label of the container.
How to mount HostPath Volume in Kubernetes with SELinux I am trying to mount a hostPath volume into a Kubernetes Pod. An example of a hostPath volume specification is shown below, which is taken from the docs. I am deploying to hosts that are running RHEL 7 with SELinux enabled. apiVersion: v1 kind: Pod metadata: name: test-pd spec: containers: - image: k8s.gcr.io/test-webserver name: test-container volumeMounts: - mountPath: /test-pd name: test-volume volumes: - name: test-volume hostPath: # directory location on host path: /data # this field is optional type: Directory When my Pod tries to read from a file that has been mounted from the underlying host, I get a "Permission Denied" error. When I run setenforce 0 to turn off SELinux, the error goes away and I can access the file. I get the same error when I bind mount a directory into a Docker container. The issue is described here and, when using Docker, can be fixed by using the z or Z bind mount flag, described in the Docker docs here . Whilst I can fix the issue by running chcon -Rt svirt_sandbox_file_t /path/to/my/host/dir/to/mount I see this as a nasty hack, as I need to do this on every host in my Kubernetes cluster and also because my deployment of Kubernetes as described in the YAML spec is not a complete description of what it is that needs to be done to get my YAML to run correctly. Turning off SELinux is not an option. I can see that Kubernetes mentions SELinux security contexts in the docs here , but I haven't been able to successfully mount a hostPath volume into a pod without getting the permission denied error. What does the YAML need to look like to successfully enable a container to mount a HostPath volume from an underlying host that is running SELinux? Update: The file I am accessing is a CA certificate that has these labels: system_u:object_r:cert_t:s0 When I use the following options: securityContext: seLinuxOptions: level: "s0:c123,c456" and then check the access control audit errors via ausearch -m avc -ts recent , I can see that there is a permission denied error where the container has a level label of s0:c123,c456 , so I can see that the level label works. I have set the label to be s0 . However, if I try to change the type label to be cert_t , the container doesn't even start, there's an error : container_linux.go:247: starting container process caused "process_linux.go:364: container init caused \"write /proc/self/task/1/attr/exec: invalid argument\"" I don't seem to be able to change the type label of the container.
docker, kubernetes, rhel, selinux
3
11,055
4
https://stackoverflow.com/questions/51000791/how-to-mount-hostpath-volume-in-kubernetes-with-selinux
24,925,332
Python HTTP SSL Server slow to accept connections
The following python code is used to create a HTTPS Server. ''' Created on 23/07/2014 @author: o880732 ''' import sys import BaseHTTPServer import base64 from multiprocessing import Process, Queue from httplib import HTTPSConnection import ssl import datetime class MyHttpEndpointHandler(BaseHTTPServer.BaseHTTPRequestHandler): def __init__(self, q, *args): self.q = q BaseHTTPServer.BaseHTTPRequestHandler.__init__(self, *args) def do_POST(self): self.send_response(200, 'OK') self.send_header('Server', 'Apache-Coyote/1.1') self.send_header('Content-type','text/xml') self.end_headers() # Extract and print the contents of the POST length = int(self.headers['Content-Length']) print "Length of content is " + str(length) post_data = self.rfile.read(length) self.q.put(post_data) class ServerEndpointHandler: def __init__(self,q,keyfile,certfile): self.q = q self.keyfile = keyfile self.certfile = certfile def setupEndpointHandler(self, queue): return lambda *args: MyHttpEndpointHandler(queue, *args) def handleEndpoint(self): print "Starting endpoint handler" handler = self.setupEndpointHandler(self.q) httpd = BaseHTTPServer.HTTPServer(('',8443), handler) httpd.socket = ssl.wrap_socket(httpd.socket, keyfile=self.keyfile, certfile = self.certfile, server_side=True) httpd.serve_forever() if __name__ == '__main__': q = Queue() print "trying to start endpoint" ep = ServerEndpointHandler(q, keyfile="~/certificates/server.key", certfile = "~/certificates/server.crt") t2 = Process(target=ep.handleEndpoint) t2.start() t2.join() The issue I have is that on a RedhatLinux server the time to accept a connection is approximately 20sec, whereas the same code running under Windows7 the connection time is approximately 200msec. Both systems are running Python 2.6.6. The redhat release is 6.4 (Santiago) Linux myserver.domain.com 2.6.32-358.18.1.el6.x86_64 #1 SMP Fri Aug 2 17:04:38 EDT 2013 x86_64 x86_64 x86_64 GNU/Linux I have updated the openssl and python packages to the current release for Redhat. Installed Packages openssl.x86_64 1.0.1e-16.el6_5.14 @rhel-x86_64-server-6 python.x86_64 2.6.6-52.el6 @rhel-x86_64-server-6 Available Packages openssl.i686 1.0.1e-16.el6_5.14 rhel-x86_64-server-6 python.i686 2.6.5-3.el6_0.2 rhel-x86_64-server-6 I am presuming that the delay has something to do with protocol negotiation and am wondering where to look to solve this issue, or what else the issue may be. Any clues?
Python HTTP SSL Server slow to accept connections The following python code is used to create a HTTPS Server. ''' Created on 23/07/2014 @author: o880732 ''' import sys import BaseHTTPServer import base64 from multiprocessing import Process, Queue from httplib import HTTPSConnection import ssl import datetime class MyHttpEndpointHandler(BaseHTTPServer.BaseHTTPRequestHandler): def __init__(self, q, *args): self.q = q BaseHTTPServer.BaseHTTPRequestHandler.__init__(self, *args) def do_POST(self): self.send_response(200, 'OK') self.send_header('Server', 'Apache-Coyote/1.1') self.send_header('Content-type','text/xml') self.end_headers() # Extract and print the contents of the POST length = int(self.headers['Content-Length']) print "Length of content is " + str(length) post_data = self.rfile.read(length) self.q.put(post_data) class ServerEndpointHandler: def __init__(self,q,keyfile,certfile): self.q = q self.keyfile = keyfile self.certfile = certfile def setupEndpointHandler(self, queue): return lambda *args: MyHttpEndpointHandler(queue, *args) def handleEndpoint(self): print "Starting endpoint handler" handler = self.setupEndpointHandler(self.q) httpd = BaseHTTPServer.HTTPServer(('',8443), handler) httpd.socket = ssl.wrap_socket(httpd.socket, keyfile=self.keyfile, certfile = self.certfile, server_side=True) httpd.serve_forever() if __name__ == '__main__': q = Queue() print "trying to start endpoint" ep = ServerEndpointHandler(q, keyfile="~/certificates/server.key", certfile = "~/certificates/server.crt") t2 = Process(target=ep.handleEndpoint) t2.start() t2.join() The issue I have is that on a RedhatLinux server the time to accept a connection is approximately 20sec, whereas the same code running under Windows7 the connection time is approximately 200msec. Both systems are running Python 2.6.6. The redhat release is 6.4 (Santiago) Linux myserver.domain.com 2.6.32-358.18.1.el6.x86_64 #1 SMP Fri Aug 2 17:04:38 EDT 2013 x86_64 x86_64 x86_64 GNU/Linux I have updated the openssl and python packages to the current release for Redhat. Installed Packages openssl.x86_64 1.0.1e-16.el6_5.14 @rhel-x86_64-server-6 python.x86_64 2.6.6-52.el6 @rhel-x86_64-server-6 Available Packages openssl.i686 1.0.1e-16.el6_5.14 rhel-x86_64-server-6 python.i686 2.6.5-3.el6_0.2 rhel-x86_64-server-6 I am presuming that the delay has something to do with protocol negotiation and am wondering where to look to solve this issue, or what else the issue may be. Any clues?
python, linux, ssl, rhel
3
970
1
https://stackoverflow.com/questions/24925332/python-http-ssl-server-slow-to-accept-connections
76,362,225
How can I resolve &#39;Unable to find a match: buildah podman&#39; error when installing on RHEL 8.7?
I am trying to install buildah and podman on my rhel 8.7 After running dnf install buildah podman -y I received this error. I tried to look for a solution on google but I am unable to find any. Please advise. [root@localhost tmp]# dnf install buildah podman -y Updating Subscription Management repositories. Last metadata expiration check: 0:14:16 ago on Mon 29 May 2023 11:32:47 PM PDT. No match for argument: buildah No match for argument: podman Error: Unable to find a match: buildah podman
How can I resolve &#39;Unable to find a match: buildah podman&#39; error when installing on RHEL 8.7? I am trying to install buildah and podman on my rhel 8.7 After running dnf install buildah podman -y I received this error. I tried to look for a solution on google but I am unable to find any. Please advise. [root@localhost tmp]# dnf install buildah podman -y Updating Subscription Management repositories. Last metadata expiration check: 0:14:16 ago on Mon 29 May 2023 11:32:47 PM PDT. No match for argument: buildah No match for argument: podman Error: Unable to find a match: buildah podman
rhel, podman, buildah
3
1,425
1
https://stackoverflow.com/questions/76362225/how-can-i-resolve-unable-to-find-a-match-buildah-podman-error-when-installing
71,516,874
Podman Non-Root &quot;Error setting up pivot dir&quot;
First time posting on StackOverflow so please be gentle! I'm setting up a new RHEL8 server to run Podman. Previously, I've done this on a pretty vanilla server but this one is setup in line with our corporate image. This means a homedir that is mounted over NFS. When I try a simple podman command such as podman run centos , I get a couple of errors (see below). According to [URL] , Podman non-root is known to have problems with NFS homedirs. Output from podman run centos (and others): ❯ podman run centos Resolved "centos" as an alias (/etc/containers/registries.conf.d/000-shortnames.conf) Trying to pull quay.io/centos/centos:latest... Getting image source signatures Copying blob 7a0437f04f83 done Error: writing blob: adding layer with blob "sha256:7a0437f04f83f084b7ed68ad9c4a4947e12fc4e1b006b38129bac89114ec3621": Error processing tar file(exit status 1): Error setting up pivot dir: mkdir /home/me/.local/share/containers/storage/overlay/2653d992f4ef2bfd27f94db643815aa567240c37732cae1405ad1c1309ee9859/diff/.pivot_root926823499: permission denied No, my username isn't really 'me' Is there a way to use podman non-root in this setup? I'd prefer to avoid creating a local user account to run things under (this is my dev server and isn't where the application will actually be running but will involve me building, running, destroying regularly so I'd rather avoid having to do anything 'clever') Output of podman info : ❯ podman info host: arch: amd64 buildahVersion: 1.23.1 cgroupControllers: [] cgroupManager: cgroupfs cgroupVersion: v1 conmon: package: conmon-2.0.32-1.module+el8.5.0+13852+150547f7.x86_64 path: /usr/bin/conmon version: 'conmon version 2.0.32, commit: 4b12bce835c3f8acc006a43620dd955a6a73bae0' cpus: 1 distribution: distribution: '"rhel"' version: "8.5" eventLogger: file hostname: servername idMappings: gidmap: - container_id: 0 host_id: 2000 size: 1 uidmap: - container_id: 0 host_id: 10279927 size: 1 kernel: 4.18.0-348.12.2.el8_5.x86_64 linkmode: dynamic logDriver: k8s-file memFree: 1881419776 memTotal: 3918233600 ociRuntime: name: runc package: runc-1.0.3-1.module+el8.5.0+13556+7f055e70.x86_64 path: /usr/bin/runc version: |- runc version 1.0.3 spec: 1.0.2-dev go: go1.16.7 libseccomp: 2.5.1 os: linux remoteSocket: path: /run/user/10279927/podman/podman.sock security: apparmorEnabled: false capabilities: CAP_NET_RAW,CAP_CHOWN,CAP_DAC_OVERRIDE,CAP_FOWNER,CAP_FSETID,CAP_KILL,CAP_NET_BIND_SERVICE,CAP_SETFCAP,CAP_SETGID,CAP_SETPCAP,CAP_SETUID,CAP_SYS_CHROOT rootless: true seccompEnabled: true seccompProfilePath: /usr/share/containers/seccomp.json selinuxEnabled: true serviceIsRemote: false slirp4netns: executable: /usr/bin/slirp4netns package: slirp4netns-1.1.8-1.module+el8.5.0+12582+56d94c81.x86_64 version: |- slirp4netns version 1.1.8 commit: d361001f495417b880f20329121e3aa431a8f90f libslirp: 4.4.0 SLIRP_CONFIG_VERSION_MAX: 3 libseccomp: 2.5.1 swapFree: 4294963200 swapTotal: 4294963200 uptime: 2h 45m 20.28s (Approximately 0.08 days) plugins: log: - k8s-file - none - journald network: - bridge - macvlan volume: - local registries: search: - registry.fedoraproject.org - registry.access.redhat.com - registry.centos.org - docker.io store: configFile: /home/me/.config/containers/storage.conf containerStore: number: 0 paused: 0 running: 0 stopped: 0 graphDriverName: overlay graphOptions: overlay.mount_program: Executable: /usr/bin/fuse-overlayfs Package: fuse-overlayfs-1.8-1.module+el8.5.0+13754+92ec836b.x86_64 Version: |- fusermount3 version: 3.2.1 fuse-overlayfs: version 1.8 FUSE library version 3.2.1 using FUSE kernel interface version 7.26 graphRoot: /home/me/.local/share/containers/storage graphStatus: Backing Filesystem: nfs Native Overlay Diff: "false" Supports d_type: "true" Using metacopy: "false" imageStore: number: 0 runRoot: /run/user/10279927/containers volumePath: /home/me/.local/share/containers/storage/volumes version: APIVersion: 3.4.2 Built: 1642068949 BuiltTime: Thu Jan 13 10:15:49 2022 GitCommit: "" GoVersion: go1.16.7 OsArch: linux/amd64 Version: 3.4.2 Thank you!
Podman Non-Root &quot;Error setting up pivot dir&quot; First time posting on StackOverflow so please be gentle! I'm setting up a new RHEL8 server to run Podman. Previously, I've done this on a pretty vanilla server but this one is setup in line with our corporate image. This means a homedir that is mounted over NFS. When I try a simple podman command such as podman run centos , I get a couple of errors (see below). According to [URL] , Podman non-root is known to have problems with NFS homedirs. Output from podman run centos (and others): ❯ podman run centos Resolved "centos" as an alias (/etc/containers/registries.conf.d/000-shortnames.conf) Trying to pull quay.io/centos/centos:latest... Getting image source signatures Copying blob 7a0437f04f83 done Error: writing blob: adding layer with blob "sha256:7a0437f04f83f084b7ed68ad9c4a4947e12fc4e1b006b38129bac89114ec3621": Error processing tar file(exit status 1): Error setting up pivot dir: mkdir /home/me/.local/share/containers/storage/overlay/2653d992f4ef2bfd27f94db643815aa567240c37732cae1405ad1c1309ee9859/diff/.pivot_root926823499: permission denied No, my username isn't really 'me' Is there a way to use podman non-root in this setup? I'd prefer to avoid creating a local user account to run things under (this is my dev server and isn't where the application will actually be running but will involve me building, running, destroying regularly so I'd rather avoid having to do anything 'clever') Output of podman info : ❯ podman info host: arch: amd64 buildahVersion: 1.23.1 cgroupControllers: [] cgroupManager: cgroupfs cgroupVersion: v1 conmon: package: conmon-2.0.32-1.module+el8.5.0+13852+150547f7.x86_64 path: /usr/bin/conmon version: 'conmon version 2.0.32, commit: 4b12bce835c3f8acc006a43620dd955a6a73bae0' cpus: 1 distribution: distribution: '"rhel"' version: "8.5" eventLogger: file hostname: servername idMappings: gidmap: - container_id: 0 host_id: 2000 size: 1 uidmap: - container_id: 0 host_id: 10279927 size: 1 kernel: 4.18.0-348.12.2.el8_5.x86_64 linkmode: dynamic logDriver: k8s-file memFree: 1881419776 memTotal: 3918233600 ociRuntime: name: runc package: runc-1.0.3-1.module+el8.5.0+13556+7f055e70.x86_64 path: /usr/bin/runc version: |- runc version 1.0.3 spec: 1.0.2-dev go: go1.16.7 libseccomp: 2.5.1 os: linux remoteSocket: path: /run/user/10279927/podman/podman.sock security: apparmorEnabled: false capabilities: CAP_NET_RAW,CAP_CHOWN,CAP_DAC_OVERRIDE,CAP_FOWNER,CAP_FSETID,CAP_KILL,CAP_NET_BIND_SERVICE,CAP_SETFCAP,CAP_SETGID,CAP_SETPCAP,CAP_SETUID,CAP_SYS_CHROOT rootless: true seccompEnabled: true seccompProfilePath: /usr/share/containers/seccomp.json selinuxEnabled: true serviceIsRemote: false slirp4netns: executable: /usr/bin/slirp4netns package: slirp4netns-1.1.8-1.module+el8.5.0+12582+56d94c81.x86_64 version: |- slirp4netns version 1.1.8 commit: d361001f495417b880f20329121e3aa431a8f90f libslirp: 4.4.0 SLIRP_CONFIG_VERSION_MAX: 3 libseccomp: 2.5.1 swapFree: 4294963200 swapTotal: 4294963200 uptime: 2h 45m 20.28s (Approximately 0.08 days) plugins: log: - k8s-file - none - journald network: - bridge - macvlan volume: - local registries: search: - registry.fedoraproject.org - registry.access.redhat.com - registry.centos.org - docker.io store: configFile: /home/me/.config/containers/storage.conf containerStore: number: 0 paused: 0 running: 0 stopped: 0 graphDriverName: overlay graphOptions: overlay.mount_program: Executable: /usr/bin/fuse-overlayfs Package: fuse-overlayfs-1.8-1.module+el8.5.0+13754+92ec836b.x86_64 Version: |- fusermount3 version: 3.2.1 fuse-overlayfs: version 1.8 FUSE library version 3.2.1 using FUSE kernel interface version 7.26 graphRoot: /home/me/.local/share/containers/storage graphStatus: Backing Filesystem: nfs Native Overlay Diff: "false" Supports d_type: "true" Using metacopy: "false" imageStore: number: 0 runRoot: /run/user/10279927/containers volumePath: /home/me/.local/share/containers/storage/volumes version: APIVersion: 3.4.2 Built: 1642068949 BuiltTime: Thu Jan 13 10:15:49 2022 GitCommit: "" GoVersion: go1.16.7 OsArch: linux/amd64 Version: 3.4.2 Thank you!
containers, rhel, podman, rhel8
3
4,289
1
https://stackoverflow.com/questions/71516874/podman-non-root-error-setting-up-pivot-dir
63,658,868
RedHat UBI 8 install expect package
I have expect script that I need to run under RedHat UBI 8 container. I'm trying to install expect via snap package manager, but I have problem starting snapd. After issuing systemctl enable --now snapd.socket I get: System has not been booted with systemd as init system (PID 1). Can't operate. Failed to connect to bus: Host is down Can be expect installed into RedHat UBI 8 without snap? My Dockerfile so far: FROM registry.access.redhat.com/ubi8/ubi RUN dnf install [URL] -y && \ dnf install [URL] -y && \ dnf update -y && \ dnf install -y wget && \ wget [URL] && \ dnf install squashfs-tools-4.0-5.el6.x86_64.rpm -y && \ wget [URL] && \ dnf install -y bash-completion-2.7-5.el8.noarch.rpm && \ dnf install snapd -y && \ systemctl enable --now snapd.socket && \ ln -s /var/lib/snapd/snap /snap && \ snap install expect
RedHat UBI 8 install expect package I have expect script that I need to run under RedHat UBI 8 container. I'm trying to install expect via snap package manager, but I have problem starting snapd. After issuing systemctl enable --now snapd.socket I get: System has not been booted with systemd as init system (PID 1). Can't operate. Failed to connect to bus: Host is down Can be expect installed into RedHat UBI 8 without snap? My Dockerfile so far: FROM registry.access.redhat.com/ubi8/ubi RUN dnf install [URL] -y && \ dnf install [URL] -y && \ dnf update -y && \ dnf install -y wget && \ wget [URL] && \ dnf install squashfs-tools-4.0-5.el6.x86_64.rpm -y && \ wget [URL] && \ dnf install -y bash-completion-2.7-5.el8.noarch.rpm && \ dnf install snapd -y && \ systemctl enable --now snapd.socket && \ ln -s /var/lib/snapd/snap /snap && \ snap install expect
docker, expect, systemd, rhel
3
15,469
2
https://stackoverflow.com/questions/63658868/redhat-ubi-8-install-expect-package
61,789,419
Write txt file in linux for .net Core (Docker)
I am new in Linux and my API was created in .net core and running in Docker. The system i create will write/create a txt file that will input all errors logged in the API. My code to write is this public class WriteLogs { public void ErrorLogFile(string traceNo, string errorMsg) { DirectoryInfo dir = new DirectoryInfo(Startup.errorPath); if (!dir.Exists) { dir.Create(); } using (StreamWriter swLog = File.AppendText(Startup.errorPath + Startup.errorFileName + DateTime.Now.ToString("MMddyyyy") + ".txt")) { swLog.WriteLine(DateTime.Now.ToString("yyyy-MM-dd hh:mm:ss.fff") + " - Trace Number : " + traceNo + " " + errorMsg + "\n"); } } } the value in my startup is located in my appSettings.json file : "ErrorPath": "C:\\BP\\", "ErrorFileName": "BP-ParamLogs_", This is working in windows environment but when i transfer my program to linux and change the ErrorPath to: "ErrorPath": "/home/Logs/", the file was not created. My question is, do my syntax works in linux to write txt file or my path was wrong?
Write txt file in linux for .net Core (Docker) I am new in Linux and my API was created in .net core and running in Docker. The system i create will write/create a txt file that will input all errors logged in the API. My code to write is this public class WriteLogs { public void ErrorLogFile(string traceNo, string errorMsg) { DirectoryInfo dir = new DirectoryInfo(Startup.errorPath); if (!dir.Exists) { dir.Create(); } using (StreamWriter swLog = File.AppendText(Startup.errorPath + Startup.errorFileName + DateTime.Now.ToString("MMddyyyy") + ".txt")) { swLog.WriteLine(DateTime.Now.ToString("yyyy-MM-dd hh:mm:ss.fff") + " - Trace Number : " + traceNo + " " + errorMsg + "\n"); } } } the value in my startup is located in my appSettings.json file : "ErrorPath": "C:\\BP\\", "ErrorFileName": "BP-ParamLogs_", This is working in windows environment but when i transfer my program to linux and change the ErrorPath to: "ErrorPath": "/home/Logs/", the file was not created. My question is, do my syntax works in linux to write txt file or my path was wrong?
linux, docker, asp.net-core, rhel, streamwriter
3
4,556
3
https://stackoverflow.com/questions/61789419/write-txt-file-in-linux-for-net-core-docker
58,481,595
How To Install Perl 5.30 in RHEL 7.7 Docker container
I want to install Perl v5.30 in my RHEL 7.7 container image. So i tried #yum install -y perl perl-5.16.3-294.el7_6.x86_64 already installed and latest version I'm able to install perl-5.16.3. But how can I install v5.30 instead of v5.16.3? I am completely new to Perl. Thanks in advance
How To Install Perl 5.30 in RHEL 7.7 Docker container I want to install Perl v5.30 in my RHEL 7.7 container image. So i tried #yum install -y perl perl-5.16.3-294.el7_6.x86_64 already installed and latest version I'm able to install perl-5.16.3. But how can I install v5.30 instead of v5.16.3? I am completely new to Perl. Thanks in advance
linux, docker, perl, installation, rhel
3
2,501
3
https://stackoverflow.com/questions/58481595/how-to-install-perl-5-30-in-rhel-7-7-docker-container
49,622,959
bash kill process if it takes too long to shutdown
I have a group of supervisord run process that I want to automate starting and stopping with a script. But sometimes the supervisor shutdown command takes too long to complete and wait times start to pile up. Currently I am doing: for each in "${args[@]}"; do (supervisorctl -c configs/supervisord_${each}.conf shutdown) & sleep 8s & (kill -9 ps aux | grep supervisor| grep ${each} | a wk '{print $2}' && kill -9 ps aux | grep supervisor| grep ${each} | awk '{print $2}') done Is there a cleaner way I can kill the supervisor process, after a few seconds, if it takes too long to stop? To be clear, I am trying to kill the supervisor process that I am trying to shutdown with supervisorctl -c {config path} shutdown , and not the supervisorctl shutdown command itself.
bash kill process if it takes too long to shutdown I have a group of supervisord run process that I want to automate starting and stopping with a script. But sometimes the supervisor shutdown command takes too long to complete and wait times start to pile up. Currently I am doing: for each in "${args[@]}"; do (supervisorctl -c configs/supervisord_${each}.conf shutdown) & sleep 8s & (kill -9 ps aux | grep supervisor| grep ${each} | a wk '{print $2}' && kill -9 ps aux | grep supervisor| grep ${each} | awk '{print $2}') done Is there a cleaner way I can kill the supervisor process, after a few seconds, if it takes too long to stop? To be clear, I am trying to kill the supervisor process that I am trying to shutdown with supervisorctl -c {config path} shutdown , and not the supervisorctl shutdown command itself.
linux, bash, centos, rhel, supervisord
3
1,765
1
https://stackoverflow.com/questions/49622959/bash-kill-process-if-it-takes-too-long-to-shutdown
34,891,668
RHEL7: How to solve “ import read failed(2)”?
I am trying to add Docker’s public key for CS packages using the following command( which is taken from the link [URL] ): sudo rpm --import "[URL] However, I am getting the following error on running the above command: curl: (6) Could not resolve host: sks-keyservers.net; Name or service not known error: [URL] import read failed(2). Is it a firewall issue?? Please let me know if I need to add a host entry in /etc/hosts ?
RHEL7: How to solve “ import read failed(2)”? I am trying to add Docker’s public key for CS packages using the following command( which is taken from the link [URL] ): sudo rpm --import "[URL] However, I am getting the following error on running the above command: curl: (6) Could not resolve host: sks-keyservers.net; Name or service not known error: [URL] import read failed(2). Is it a firewall issue?? Please let me know if I need to add a host entry in /etc/hosts ?
linux, docker, centos, rhel, docker-registry
3
14,802
1
https://stackoverflow.com/questions/34891668/rhel7-how-to-solve-import-read-failed2
10,625,267
Approach for installing system service implemented as Ruby gem
After years of being away from Ruby, I'm back full-force and have just cut my first gem, which includes an executable. Everything works like a charm. The problem I am facing, however, is that I ALSO have a startup script (not part of the gem istelf) that daemonizes the executable. Additionally, I'd also like for the startup script to point the executable at configuration in a place like /var/ To the best of my knowledge, there's no way with rubygems, gemspec, etc., to specify files getting blown out to other parts of your system during install (e.g. startup script to /etc/init.d, and config to /var/). It certainly wouldn't make sense if you COULD do that. So... my question is... what IS the proper procedure for automating the installation of something like this. I'm using RHEL, and am wondering if it's, perhaps, time for me to get my feet wet with making my first RPM. Any thoughts?
Approach for installing system service implemented as Ruby gem After years of being away from Ruby, I'm back full-force and have just cut my first gem, which includes an executable. Everything works like a charm. The problem I am facing, however, is that I ALSO have a startup script (not part of the gem istelf) that daemonizes the executable. Additionally, I'd also like for the startup script to point the executable at configuration in a place like /var/ To the best of my knowledge, there's no way with rubygems, gemspec, etc., to specify files getting blown out to other parts of your system during install (e.g. startup script to /etc/init.d, and config to /var/). It certainly wouldn't make sense if you COULD do that. So... my question is... what IS the proper procedure for automating the installation of something like this. I'm using RHEL, and am wondering if it's, perhaps, time for me to get my feet wet with making my first RPM. Any thoughts?
ruby, linux, installation, daemon, rhel
3
318
1
https://stackoverflow.com/questions/10625267/approach-for-installing-system-service-implemented-as-ruby-gem
74,527,160
How to fix &quot;An error ocurred while loading the archive.&quot; in Archive Manager in RHEL 7.6 when opening ZIP files?
When I try to open a ZIP file in the Archive Manager, I get this error: "An error ocurred while loading the archive." This issue is happening on a virtual machine running Red Hat Enterprise Linux 7.6. Oddly, this did not happen to me prior to this year 2022. Things worked OK in 2021 and each year before. How do I fix this error "An error ocurred while loading the archive?" I reverted to an old snapshot of my RHEL 7.6 virtual machine and I can now open ZIP files in the Archive Manager with zero errors. I think a system setting was edited OR a newly installed program is causing the issue.
How to fix &quot;An error ocurred while loading the archive.&quot; in Archive Manager in RHEL 7.6 when opening ZIP files? When I try to open a ZIP file in the Archive Manager, I get this error: "An error ocurred while loading the archive." This issue is happening on a virtual machine running Red Hat Enterprise Linux 7.6. Oddly, this did not happen to me prior to this year 2022. Things worked OK in 2021 and each year before. How do I fix this error "An error ocurred while loading the archive?" I reverted to an old snapshot of my RHEL 7.6 virtual machine and I can now open ZIP files in the Archive Manager with zero errors. I think a system setting was edited OR a newly installed program is causing the issue.
zip, rhel
3
7,050
1
https://stackoverflow.com/questions/74527160/how-to-fix-an-error-ocurred-while-loading-the-archive-in-archive-manager-in-r
51,981,132
RHEL7 import tkinter failed inside virtualenv
On Redhat 7 and python3.6 I can import tkinter without any error: Python 3.6.5 (default, Jul 25 2018, 21:22:33) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import tkinter >>> Within a fresh virtualenv Python 3.6.4 (default, Jul 20 2018, 12:22:32) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import tkinter Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/ocr1/CRNN_Tensorflow/tkinter/__init__.py", line 36, in <module> import _tkinter # If this fails your Python may not be configured for Tk ModuleNotFoundError: No module named '_tkinter' >>> P.S. It's an entire new virtualenv, created by virtualenv <name_of_env> command.
RHEL7 import tkinter failed inside virtualenv On Redhat 7 and python3.6 I can import tkinter without any error: Python 3.6.5 (default, Jul 25 2018, 21:22:33) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import tkinter >>> Within a fresh virtualenv Python 3.6.4 (default, Jul 20 2018, 12:22:32) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import tkinter Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/ocr1/CRNN_Tensorflow/tkinter/__init__.py", line 36, in <module> import _tkinter # If this fails your Python may not be configured for Tk ModuleNotFoundError: No module named '_tkinter' >>> P.S. It's an entire new virtualenv, created by virtualenv <name_of_env> command.
python-3.x, tkinter, virtualenv, rhel
3
426
2
https://stackoverflow.com/questions/51981132/rhel7-import-tkinter-failed-inside-virtualenv
45,824,178
Install Kubernetes on Red Hat Linux (RHEL 7)
I am trying to install Kubernetes on Red Hat Linux (RHEL 7) . Any advice on the best and easiest way to do this ? I would not like to use minikube. Thank you very much
Install Kubernetes on Red Hat Linux (RHEL 7) I am trying to install Kubernetes on Red Hat Linux (RHEL 7) . Any advice on the best and easiest way to do this ? I would not like to use minikube. Thank you very much
kubernetes, rhel
3
5,782
4
https://stackoverflow.com/questions/45824178/install-kubernetes-on-red-hat-linux-rhel-7
42,126,502
No tkinter package available
While trying to import matplotlib on an RHEL Fedora system, I get this error: Traceback (most recent call last): File "generate_example_images.py", line 9, in <module> import matplotlib.pyplot as plt File "/usr/local/lib64/python2.7/site-packages/matplotlib/pyplot.py", line 115, in <module> _backend_mod, new_figure_manager, draw_if_interactive, _show = pylab_setup() File "/usr/local/lib64/python2.7/site-packages/matplotlib/backends/__init__.py", line 32, in pylab_setup globals(),locals(),[backend_name],0) File "/usr/local/lib64/python2.7/site-packages/matplotlib/backends/backend_tkagg.py", line 6, in <module> from six.moves import tkinter as Tk File "/usr/lib/python2.7/dist-packages/six.py", line 194, in load_module mod = mod._resolve() File "/usr/lib/python2.7/dist-packages/six.py", line 108, in _resolve return _import_module(self.mod) File "/usr/lib/python2.7/dist-packages/six.py", line 79, in _import_module __import__(name) ImportError: No module named Tkinter Another question at [URL] says I can use the following: yum install tkinter This should install a tkinter package for python 2 and python3, however, I then get the error Loaded plugins: priorities, update-motd, upgrade-helper amzn-main/latest | 2.1 kB 00:00 amzn-updates/latest | 2.3 kB 00:00 No package tkinter available. Error: Nothing to do when I look at the yum packages I also don't see a tkinter package so it makes sense I'm getting this error.
No tkinter package available While trying to import matplotlib on an RHEL Fedora system, I get this error: Traceback (most recent call last): File "generate_example_images.py", line 9, in <module> import matplotlib.pyplot as plt File "/usr/local/lib64/python2.7/site-packages/matplotlib/pyplot.py", line 115, in <module> _backend_mod, new_figure_manager, draw_if_interactive, _show = pylab_setup() File "/usr/local/lib64/python2.7/site-packages/matplotlib/backends/__init__.py", line 32, in pylab_setup globals(),locals(),[backend_name],0) File "/usr/local/lib64/python2.7/site-packages/matplotlib/backends/backend_tkagg.py", line 6, in <module> from six.moves import tkinter as Tk File "/usr/lib/python2.7/dist-packages/six.py", line 194, in load_module mod = mod._resolve() File "/usr/lib/python2.7/dist-packages/six.py", line 108, in _resolve return _import_module(self.mod) File "/usr/lib/python2.7/dist-packages/six.py", line 79, in _import_module __import__(name) ImportError: No module named Tkinter Another question at [URL] says I can use the following: yum install tkinter This should install a tkinter package for python 2 and python3, however, I then get the error Loaded plugins: priorities, update-motd, upgrade-helper amzn-main/latest | 2.1 kB 00:00 amzn-updates/latest | 2.3 kB 00:00 No package tkinter available. Error: Nothing to do when I look at the yum packages I also don't see a tkinter package so it makes sense I'm getting this error.
python, tkinter, rhel
3
5,249
1
https://stackoverflow.com/questions/42126502/no-tkinter-package-available
41,573,690
Docker image just contain software without OS
I have a PROD environment running on RHEL 7 server. I want to use docker for deployment. I want to package all the software and apps in a Docker image, without a base OS. Because I don't want to add an additional layer on top of RHEL. Also, I could not find an official base image for RHEL. Is that possible? I see some old posts mentioned about "FROM scratch" but looks it does not work in the latest version of Docker -- 1.12.5. If this is impossible, any suggestions for this?
Docker image just contain software without OS I have a PROD environment running on RHEL 7 server. I want to use docker for deployment. I want to package all the software and apps in a Docker image, without a base OS. Because I don't want to add an additional layer on top of RHEL. Also, I could not find an official base image for RHEL. Is that possible? I see some old posts mentioned about "FROM scratch" but looks it does not work in the latest version of Docker -- 1.12.5. If this is impossible, any suggestions for this?
docker, dockerfile, rhel
3
1,839
1
https://stackoverflow.com/questions/41573690/docker-image-just-contain-software-without-os
39,714,070
BCP neither gives results nor outputs anything when using valid statements but it does throw errors when passing invalid parameters
I have to use bcp command-line tool to export data from an SQL Server database to a file in a Red Hat server. I am (apparently) using valid statements but bcp is not producing any kind of output/results. However, when I execute statements with missing or invalid parameters it displays the respective error. I am looking for the reason of this issue (e.g. defective installation, bad usage of bcp , lack of permissions or any other known conflict) and how to fix it. bcp statement: bcp fully_qualified_table_name out ./data.txt -c -S server -U user -P password bcp usage: usage: /opt/microsoft/bin/bcp {dbtable | query} {in | out | queryout | format} datafile [-m maxerrors] [-f formatfile] [-e errfile] [-F firstrow] [-L lastrow] [-b batchsize] [-n native type] [-c character type] [-w wide character type] [-N keep non-text native] [-q quoted identifier] [-t field terminator] [-r row terminator] [-a packetsize] [-K application intent] [-S server name or DSN if -D provided] [-D treat -S as DSN] [-U username] [-P password] [-T trusted connection] [-v version] [-R regional enable] [-k keep null values] [-E keep identity values] [-h "load hints"] [-d database name] bcp version: BCP - Bulk Copy Program for Microsoft SQL Server. Copyright (C) Microsoft Corporation. All Rights Reserved. Version: 11.0.2270.0 SQL Server version ( SELECT @@VERSION ): Microsoft SQL Server 2012 - 11.0.5058.0 (X64) May 14 2014 18:34:29 Copyright (c) Microsoft Corporation Enterprise Edition: Core-based Licensing (64-bit) on Windows NT 6.3 <X64> (Build 9600: ) (Hypervisor) Distribution: Red Hat Enterprise Linux 6.7 (KornShell). Invalid statements with respective error message (examples). bcp THAT_TUB_ACE.oh_nerd.table_name out ./data.txt -c -S sr._bear -U you_sr. -P pass_sword SQLState = S1T00, NativeError = 0 Error = [unixODBC][Microsoft][ODBC Driver 11 for SQL Server]Login timeout expired SQLState = 08001, NativeError = 11001 Error = [unixODBC][Microsoft][ODBC Driver 11 for SQL Server]A network-related or instance-specific error has occurred while establishing a connection to SQL Server. Server is not found or not accessible. Check if instance name is correct and if SQL Server is configured to allow remote connections. For more information see SQL Server Books Online. SQLState = 08001, NativeError = 11001 Error = [unixODBC][Microsoft][ODBC Driver 11 for SQL Server]TCP Provider: Error code 0x2AF9 ... bcp fully_qualified_table_name ./data.txt -c -S valid_server -U valid_user -P bad_word bcp fully_qualified_table_name out ./data.txt -c -S valid_server -U valid_user -P bad_word SQLState = 28000, NativeError = 18456 Error = [unixODBC][Microsoft][ODBC Driver 11 for SQL Server][SQL Server]Login failed for user 'valid_user'. SUMMARY. The objective is to generate a datafile using the following syntax (or similar): bcp fully_qualified_table_name out ./data.txt -c -S server -U user -P password The facts are: When running a valid bcp statement there's nothing in the window at all (no output) and no datafile is created. I cannot use option -T (trusted connection using integrated security) for bcp so I have to specify the server, user and password . Tried queryout option on a very simple small table already but still no luck. Credentials are valid, I successfully tested them using sqlcmd like the following: sqlcmd -S server -U user -P password -Q 'SELECT * FROM really_small_table' . The bcp statements under "Invalid statements with respective error message (examples)" section of this question are just examples of invalid statements to show that bcp actually does something but giving the expected results.
BCP neither gives results nor outputs anything when using valid statements but it does throw errors when passing invalid parameters I have to use bcp command-line tool to export data from an SQL Server database to a file in a Red Hat server. I am (apparently) using valid statements but bcp is not producing any kind of output/results. However, when I execute statements with missing or invalid parameters it displays the respective error. I am looking for the reason of this issue (e.g. defective installation, bad usage of bcp , lack of permissions or any other known conflict) and how to fix it. bcp statement: bcp fully_qualified_table_name out ./data.txt -c -S server -U user -P password bcp usage: usage: /opt/microsoft/bin/bcp {dbtable | query} {in | out | queryout | format} datafile [-m maxerrors] [-f formatfile] [-e errfile] [-F firstrow] [-L lastrow] [-b batchsize] [-n native type] [-c character type] [-w wide character type] [-N keep non-text native] [-q quoted identifier] [-t field terminator] [-r row terminator] [-a packetsize] [-K application intent] [-S server name or DSN if -D provided] [-D treat -S as DSN] [-U username] [-P password] [-T trusted connection] [-v version] [-R regional enable] [-k keep null values] [-E keep identity values] [-h "load hints"] [-d database name] bcp version: BCP - Bulk Copy Program for Microsoft SQL Server. Copyright (C) Microsoft Corporation. All Rights Reserved. Version: 11.0.2270.0 SQL Server version ( SELECT @@VERSION ): Microsoft SQL Server 2012 - 11.0.5058.0 (X64) May 14 2014 18:34:29 Copyright (c) Microsoft Corporation Enterprise Edition: Core-based Licensing (64-bit) on Windows NT 6.3 <X64> (Build 9600: ) (Hypervisor) Distribution: Red Hat Enterprise Linux 6.7 (KornShell). Invalid statements with respective error message (examples). bcp THAT_TUB_ACE.oh_nerd.table_name out ./data.txt -c -S sr._bear -U you_sr. -P pass_sword SQLState = S1T00, NativeError = 0 Error = [unixODBC][Microsoft][ODBC Driver 11 for SQL Server]Login timeout expired SQLState = 08001, NativeError = 11001 Error = [unixODBC][Microsoft][ODBC Driver 11 for SQL Server]A network-related or instance-specific error has occurred while establishing a connection to SQL Server. Server is not found or not accessible. Check if instance name is correct and if SQL Server is configured to allow remote connections. For more information see SQL Server Books Online. SQLState = 08001, NativeError = 11001 Error = [unixODBC][Microsoft][ODBC Driver 11 for SQL Server]TCP Provider: Error code 0x2AF9 ... bcp fully_qualified_table_name ./data.txt -c -S valid_server -U valid_user -P bad_word bcp fully_qualified_table_name out ./data.txt -c -S valid_server -U valid_user -P bad_word SQLState = 28000, NativeError = 18456 Error = [unixODBC][Microsoft][ODBC Driver 11 for SQL Server][SQL Server]Login failed for user 'valid_user'. SUMMARY. The objective is to generate a datafile using the following syntax (or similar): bcp fully_qualified_table_name out ./data.txt -c -S server -U user -P password The facts are: When running a valid bcp statement there's nothing in the window at all (no output) and no datafile is created. I cannot use option -T (trusted connection using integrated security) for bcp so I have to specify the server, user and password . Tried queryout option on a very simple small table already but still no luck. Credentials are valid, I successfully tested them using sqlcmd like the following: sqlcmd -S server -U user -P password -Q 'SELECT * FROM really_small_table' . The bcp statements under "Invalid statements with respective error message (examples)" section of this question are just examples of invalid statements to show that bcp actually does something but giving the expected results.
sql-server, linux, database, rhel, bcp
3
1,440
1
https://stackoverflow.com/questions/39714070/bcp-neither-gives-results-nor-outputs-anything-when-using-valid-statements-but-i
35,434,947
Integrating RHEL and Active Directory
I'm integrating RHEL 6.5 server and Windows Server 2008 R2 (Active Directory) for authentication. i.e Clients are able to login into RHEL serve using user details in AD. I followed same procedure(Configuration 3) [URL] and [URL] I configured same and it was not working as expected. I'm getting a error like debug1: Authentications that can continue: publickey,gssapi-keyex,gssapi-with-mic,password debug1: Next authentication method: gssapi-keyex debug1: No valid Key exchange context debug1: Next authentication method: gssapi-with-mic debug1: Unspecified GSS failure. Minor code may provide more information Server not found in Kerberos database debug1: Unspecified GSS failure. Minor code may provide more information Server not found in Kerberos database debug1: Unspecified GSS failure. Minor code may provide more information Could you help me on this?
Integrating RHEL and Active Directory I'm integrating RHEL 6.5 server and Windows Server 2008 R2 (Active Directory) for authentication. i.e Clients are able to login into RHEL serve using user details in AD. I followed same procedure(Configuration 3) [URL] and [URL] I configured same and it was not working as expected. I'm getting a error like debug1: Authentications that can continue: publickey,gssapi-keyex,gssapi-with-mic,password debug1: Next authentication method: gssapi-keyex debug1: No valid Key exchange context debug1: Next authentication method: gssapi-with-mic debug1: Unspecified GSS failure. Minor code may provide more information Server not found in Kerberos database debug1: Unspecified GSS failure. Minor code may provide more information Server not found in Kerberos database debug1: Unspecified GSS failure. Minor code may provide more information Could you help me on this?
active-directory, rhel
3
1,215
1
https://stackoverflow.com/questions/35434947/integrating-rhel-and-active-directory
17,115,340
RHEL service wouldn&#39;t start using Puppet
I was trying to set up gitlab using this module ( [URL] ) with some modification to work with RVM and RHEL 6. I got everything working except the gitlab service wouldn't start. The service is defined as: service { 'gitlab': ensure => running, pattern => 'puma', hasrestart => true, enable => true, hasstatus => false, subscribe => File['/etc/init.d/gitlab'], } The related output from agent with debug enabled: Debug: Service[gitlab](provider=redhat): Executing 'ps -ef' Debug: Executing '/sbin/chkconfig gitlab' Debug: Executing '/sbin/service gitlab start' Debug: Executing '/sbin/chkconfig gitlab' Notice: /Stage[main]/Gitlab::Server/Service[gitlab]/ensure: ensure changed 'stopped' to 'running' Debug: /Stage[main]/Gitlab::Server/Service[gitlab]: The container Class[Gitlab::Server] will propagate my refresh even t Info: /Stage[main]/Gitlab::Server/Service[gitlab]: Unscheduling refresh on Service[gitlab] Debug: Class[Gitlab::Server]: The container Stage[main] will propagate my refresh event Debug: Finishing transaction 70257873749220 It seems the /sbin/service gitlab start got executed. But the service is not started (no entries in log either). However, if I manually run service gitlab start, it works fine. I noticed that there is a line saying "Unscheduling refresh on Service[gitlab]". Not sure why and if it has anything to do with it. Any advice? Thanks.
RHEL service wouldn&#39;t start using Puppet I was trying to set up gitlab using this module ( [URL] ) with some modification to work with RVM and RHEL 6. I got everything working except the gitlab service wouldn't start. The service is defined as: service { 'gitlab': ensure => running, pattern => 'puma', hasrestart => true, enable => true, hasstatus => false, subscribe => File['/etc/init.d/gitlab'], } The related output from agent with debug enabled: Debug: Service[gitlab](provider=redhat): Executing 'ps -ef' Debug: Executing '/sbin/chkconfig gitlab' Debug: Executing '/sbin/service gitlab start' Debug: Executing '/sbin/chkconfig gitlab' Notice: /Stage[main]/Gitlab::Server/Service[gitlab]/ensure: ensure changed 'stopped' to 'running' Debug: /Stage[main]/Gitlab::Server/Service[gitlab]: The container Class[Gitlab::Server] will propagate my refresh even t Info: /Stage[main]/Gitlab::Server/Service[gitlab]: Unscheduling refresh on Service[gitlab] Debug: Class[Gitlab::Server]: The container Stage[main] will propagate my refresh event Debug: Finishing transaction 70257873749220 It seems the /sbin/service gitlab start got executed. But the service is not started (no entries in log either). However, if I manually run service gitlab start, it works fine. I noticed that there is a line saying "Unscheduling refresh on Service[gitlab]". Not sure why and if it has anything to do with it. Any advice? Thanks.
puppet, rhel, gitlab
3
2,340
1
https://stackoverflow.com/questions/17115340/rhel-service-wouldnt-start-using-puppet
16,395,082
python daemon killed (by kernel?)
I have a python2.7 daemon process using a module from [URL] The process is a heavy one with about 40 GB RAM usage and 9 child threads. Server uses RHEL 6.3 with 192 GB RAM and enough CPU power. After starting the process, it lasts for around 3-7 hours, but then it was killed by someone, might be the kernel. But I could not find any hints in dmesg nor kernel log (which I had manually activated), nothings there. When not starting as daemon, I just got the message in terminal: "killed". The following precautions have been done: resetting the oom score in /proc//oom_score_adj so that the oom killer does not pick the process when sort of resources increasing all rlimits (that can be increased) to maximum set the process nice/priority higher (prio -15) This problem exists already before applying these precautions, so they are not responsible for the killing I also have a mechanism to catch all exception, STDERR, STDOUT and log everythings into a rotated log file. But there was nothing interesting just before the process died. Modules used within the process among others: oracle_cx, ibm_db, suds, wsgi_utils. But all of them always write logs when errors occured. Anyone know how to trace back the killing? Who and why? Thank you in advance
python daemon killed (by kernel?) I have a python2.7 daemon process using a module from [URL] The process is a heavy one with about 40 GB RAM usage and 9 child threads. Server uses RHEL 6.3 with 192 GB RAM and enough CPU power. After starting the process, it lasts for around 3-7 hours, but then it was killed by someone, might be the kernel. But I could not find any hints in dmesg nor kernel log (which I had manually activated), nothings there. When not starting as daemon, I just got the message in terminal: "killed". The following precautions have been done: resetting the oom score in /proc//oom_score_adj so that the oom killer does not pick the process when sort of resources increasing all rlimits (that can be increased) to maximum set the process nice/priority higher (prio -15) This problem exists already before applying these precautions, so they are not responsible for the killing I also have a mechanism to catch all exception, STDERR, STDOUT and log everythings into a rotated log file. But there was nothing interesting just before the process died. Modules used within the process among others: oracle_cx, ibm_db, suds, wsgi_utils. But all of them always write logs when errors occured. Anyone know how to trace back the killing? Who and why? Thank you in advance
python, linux, daemon, rhel
3
1,829
2
https://stackoverflow.com/questions/16395082/python-daemon-killed-by-kernel
8,653,531
Installing/building git-svn for RHEL 6
I do not have access to the RHEL6 repo, so what wouold be the best approach to building and/or installing the git-svn tool on RedHat Enterprise Linux 6? Managed to install git & git-svn 1.7.9 by uninstall the existing one (yum erase git) and make installing the latest one from source. Worked on 64-bit too.
Installing/building git-svn for RHEL 6 I do not have access to the RHEL6 repo, so what wouold be the best approach to building and/or installing the git-svn tool on RedHat Enterprise Linux 6? Managed to install git & git-svn 1.7.9 by uninstall the existing one (yum erase git) and make installing the latest one from source. Worked on 64-bit too.
git, git-svn, rhel
3
6,114
1
https://stackoverflow.com/questions/8653531/installing-building-git-svn-for-rhel-6
78,109,685
Jenkins Fontconfig head is null even after installing fontconfig
I am trying to install generic Jenkins with jenkins.war. Getting following error while starting. 2024-03-05 15:23:36.018+0000 [id=1] SEVERE hudson.util.BootFailure#publish: Failed to initialize Jenkins java.lang.RuntimeException: Fontconfig head is null, check your fonts or fonts configuration at java.desktop/sun.awt.FontConfiguration.getVersion(FontConfiguration.java:1263) I have installed fontconfig rpm -qa | grep fontconfig fontconfig-2.8.0-5.el6.x86_64 fontconfig-devel-2.8.0-5.el6.x86_64 I also tried starting with -Djava.awt.headless=true I am on openjdk 21 (tried both Eclipse and Openjdk.net dist) Jenkins 2.440.1 Red Hat Enterprise Linux Server release 6.10 (Santiago) base-4.0-amd64:base-4.0-noarch:core-4.0-amd64:core-4.0-noarch
Jenkins Fontconfig head is null even after installing fontconfig I am trying to install generic Jenkins with jenkins.war. Getting following error while starting. 2024-03-05 15:23:36.018+0000 [id=1] SEVERE hudson.util.BootFailure#publish: Failed to initialize Jenkins java.lang.RuntimeException: Fontconfig head is null, check your fonts or fonts configuration at java.desktop/sun.awt.FontConfiguration.getVersion(FontConfiguration.java:1263) I have installed fontconfig rpm -qa | grep fontconfig fontconfig-2.8.0-5.el6.x86_64 fontconfig-devel-2.8.0-5.el6.x86_64 I also tried starting with -Djava.awt.headless=true I am on openjdk 21 (tried both Eclipse and Openjdk.net dist) Jenkins 2.440.1 Red Hat Enterprise Linux Server release 6.10 (Santiago) base-4.0-amd64:base-4.0-noarch:core-4.0-amd64:core-4.0-noarch
jenkins, rhel
3
1,284
0
https://stackoverflow.com/questions/78109685/jenkins-fontconfig-head-is-null-even-after-installing-fontconfig
68,662,267
&#39;docker stop&#39; impact on established tcp connection
I'm running 2 docker containers(TcpServer,TcpClient). In each container, there's an init.sh script which launches the applications. In the init.sh script, I've handled SIGTERM but I'm not doing any sort of handling for that(I'm not passing it to my application). trap 'true' SIGTERM After startup, a tcp connection is established between TcpServer and TcpClient. TcpClient is a multi-threaded application, with 1 thread(receiver) doing: while(true) { //blocking tcp receive function call //process received data or received error code. } So basically, the idea is that the receiver thread would always get to know about server going down 'cleanly'. The observation is that, most of the times, when I issue 'docker stop serverContainer', the client application receives TCP 'FIN' packet after about 10 seconds. This is as per my expectations because docker first tries to kill via SIGTERM but since that is handled it then issues SIGKILL which it does only after about 10 seconds. My current understanding is that, whenever sigkill/unhandled-sigterm is given to a process, the kernel will terminate that process and close all file descriptors opened by that process. If this is true, then I should always see a FIN packet going from server to client as soon as the process is killed. However, a few times, FIN packet is not observed in the traces captured on both client and server end. As a result, the client doesn't get to know about the server getting down for a longer time(until it tries to send some data on that connection or the TCP's keepalive mechanism kicks in). I'm not sure how this happens because if I explicitly issue SIGKILL to pid 1 of my server's container(from outside), then I've always seen the FIN packet. So why sometimes, and only when using docker stop, does server not send TCP FIN? Basically I want to ask 2 things: In Linux, when SIGKILL is issued to a TCP server process, is it guaranteed that the kernel/tcp stack will send TCP FIN packet to client before terminating? When I use 'docker stop' how exactly are the processes spawned by the main process(PID 1 inside container) terminated? Because from what I had read, the SIGTERM/SIGKILL is given only to PID 1? So why are its child processes not adopted by init/systemd as happens otherwise(killing the parent process created outside the container). Operating System: Red Hat Enterprise Linux Server release 7.7 (Maipo) Docker version: Docker version 19.03.14, build 5eb3275d40
&#39;docker stop&#39; impact on established tcp connection I'm running 2 docker containers(TcpServer,TcpClient). In each container, there's an init.sh script which launches the applications. In the init.sh script, I've handled SIGTERM but I'm not doing any sort of handling for that(I'm not passing it to my application). trap 'true' SIGTERM After startup, a tcp connection is established between TcpServer and TcpClient. TcpClient is a multi-threaded application, with 1 thread(receiver) doing: while(true) { //blocking tcp receive function call //process received data or received error code. } So basically, the idea is that the receiver thread would always get to know about server going down 'cleanly'. The observation is that, most of the times, when I issue 'docker stop serverContainer', the client application receives TCP 'FIN' packet after about 10 seconds. This is as per my expectations because docker first tries to kill via SIGTERM but since that is handled it then issues SIGKILL which it does only after about 10 seconds. My current understanding is that, whenever sigkill/unhandled-sigterm is given to a process, the kernel will terminate that process and close all file descriptors opened by that process. If this is true, then I should always see a FIN packet going from server to client as soon as the process is killed. However, a few times, FIN packet is not observed in the traces captured on both client and server end. As a result, the client doesn't get to know about the server getting down for a longer time(until it tries to send some data on that connection or the TCP's keepalive mechanism kicks in). I'm not sure how this happens because if I explicitly issue SIGKILL to pid 1 of my server's container(from outside), then I've always seen the FIN packet. So why sometimes, and only when using docker stop, does server not send TCP FIN? Basically I want to ask 2 things: In Linux, when SIGKILL is issued to a TCP server process, is it guaranteed that the kernel/tcp stack will send TCP FIN packet to client before terminating? When I use 'docker stop' how exactly are the processes spawned by the main process(PID 1 inside container) terminated? Because from what I had read, the SIGTERM/SIGKILL is given only to PID 1? So why are its child processes not adopted by init/systemd as happens otherwise(killing the parent process created outside the container). Operating System: Red Hat Enterprise Linux Server release 7.7 (Maipo) Docker version: Docker version 19.03.14, build 5eb3275d40
docker, tcp, linux-kernel, network-programming, rhel
3
516
0
https://stackoverflow.com/questions/68662267/docker-stop-impact-on-established-tcp-connection
67,989,386
Segmentation fault using devtoolset-10 with boost 1.73 static libraries complied with gcc 7.5.0
I have boost 1.73 static libraries (.a) and RHEL devtoolset-10 installed,I do the following steps scl enable devtoolset-10 bash Check the LD_LIBRARY_PATH points to /opt/rh/devtoolset-10/root/usr/lib64:/opt/rh/devtoolset-10/root/usr/lib:/opt/rh/devtoolset-10/root/usr/lib64/dyninst:/opt/rh/devtoolset-10/root/usr/lib/dyninst:/opt/rh/devtoolset-10/root/usr/lib64:/opt/rh/devtoolset-10/root/usr/lib I set the CXX and CC to /opt/rh/devtoolset-10/root/bin/g++ and /opt/rh/devtoolset-10/root/bin/gcc respectively This is snippet of my makefile LPATH := -L/home/boost/lib/linux.2_6.x86_64 \ -L/opt/rh/devtoolset-10/root/bin/ \ -L/opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10 \ -L/opt/rh/devtoolset-10/root/usr/lib64 LFLAG := -lboost_filesystem -lboost_system -l:libstdc++.a -l:libgcc.a -static .PHONY: all clean all: $(EXE) $(EXE): $(BUILTOBJ) $(CXX) -DBOOST_NO_CXX11_SCOPED_ENUMS -Xlinker -Map=rhel.map $(LPATH) -o $@ $(BUILTOBJ) $(LFLAG) $(OBJ_DIR)/%.o: $(SRC_DIR)/%.cpp $(CXX) -DBOOST_NO_CXX11_SCOPED_ENUMS $(INCLUDES) -c $< -o $@ $(OBJ_DIR)/%.o: %.c mkdir -p $(@D) $(CC) $(INCLUDES) -c $< -o $@ Segmentation fault occurs at these three places..I got the boost functions that throw error ,these are boost::filesystem::path src(sourcepath); src.filename() for(boost::filesystem::directory_iterator file(src); file!=boost::filesystem::directory_iterator(); ++file) boost::filesystem::remove_all(sourcepath); Also a snippet on my rhel.map file (boost::filesystem::path::operator/=(boost::filesystem::path const&)) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(bad_alloc.o) /home/boost/lib/linux.2_6.x86_64/libboost_filesystem.a(directory.o) (typeinfo for std::bad_alloc) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(class_type_info.o) obj/Action.o (vtable for __cxxabiv1::__class_type_info) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(del_op.o) /home/boost/lib/linux.2_6.x86_64/libboost_filesystem.a(exception.o) (operator delete(void*)) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(del_ops.o) obj/Action.o (operator delete(void*, unsigned long)) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(del_opv.o) obj/CZip.o (operator delete) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(dyncast.o) obj/Action.o (__dynamic_cast) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(eh_alloc.o) obj/Action.o (__cxa_allocate_exception) Also if I use the -g flag I get this and similar DWARF debug error DWARF error: could not find variable specification at offset 23b28 The code works fine in ubuntu which has gcc 9.3.1 but in rhel-7 with devtoolset it gives segmentation fault(core dumped) Is there some error with my linking?
Segmentation fault using devtoolset-10 with boost 1.73 static libraries complied with gcc 7.5.0 I have boost 1.73 static libraries (.a) and RHEL devtoolset-10 installed,I do the following steps scl enable devtoolset-10 bash Check the LD_LIBRARY_PATH points to /opt/rh/devtoolset-10/root/usr/lib64:/opt/rh/devtoolset-10/root/usr/lib:/opt/rh/devtoolset-10/root/usr/lib64/dyninst:/opt/rh/devtoolset-10/root/usr/lib/dyninst:/opt/rh/devtoolset-10/root/usr/lib64:/opt/rh/devtoolset-10/root/usr/lib I set the CXX and CC to /opt/rh/devtoolset-10/root/bin/g++ and /opt/rh/devtoolset-10/root/bin/gcc respectively This is snippet of my makefile LPATH := -L/home/boost/lib/linux.2_6.x86_64 \ -L/opt/rh/devtoolset-10/root/bin/ \ -L/opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10 \ -L/opt/rh/devtoolset-10/root/usr/lib64 LFLAG := -lboost_filesystem -lboost_system -l:libstdc++.a -l:libgcc.a -static .PHONY: all clean all: $(EXE) $(EXE): $(BUILTOBJ) $(CXX) -DBOOST_NO_CXX11_SCOPED_ENUMS -Xlinker -Map=rhel.map $(LPATH) -o $@ $(BUILTOBJ) $(LFLAG) $(OBJ_DIR)/%.o: $(SRC_DIR)/%.cpp $(CXX) -DBOOST_NO_CXX11_SCOPED_ENUMS $(INCLUDES) -c $< -o $@ $(OBJ_DIR)/%.o: %.c mkdir -p $(@D) $(CC) $(INCLUDES) -c $< -o $@ Segmentation fault occurs at these three places..I got the boost functions that throw error ,these are boost::filesystem::path src(sourcepath); src.filename() for(boost::filesystem::directory_iterator file(src); file!=boost::filesystem::directory_iterator(); ++file) boost::filesystem::remove_all(sourcepath); Also a snippet on my rhel.map file (boost::filesystem::path::operator/=(boost::filesystem::path const&)) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(bad_alloc.o) /home/boost/lib/linux.2_6.x86_64/libboost_filesystem.a(directory.o) (typeinfo for std::bad_alloc) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(class_type_info.o) obj/Action.o (vtable for __cxxabiv1::__class_type_info) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(del_op.o) /home/boost/lib/linux.2_6.x86_64/libboost_filesystem.a(exception.o) (operator delete(void*)) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(del_ops.o) obj/Action.o (operator delete(void*, unsigned long)) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(del_opv.o) obj/CZip.o (operator delete) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(dyncast.o) obj/Action.o (__dynamic_cast) /opt/rh/devtoolset-10/root/usr/lib/gcc/x86_64-redhat-linux/10/libstdc++.a(eh_alloc.o) obj/Action.o (__cxa_allocate_exception) Also if I use the -g flag I get this and similar DWARF debug error DWARF error: could not find variable specification at offset 23b28 The code works fine in ubuntu which has gcc 9.3.1 but in rhel-7 with devtoolset it gives segmentation fault(core dumped) Is there some error with my linking?
boost, g++, static-libraries, rhel, devtoolset
3
293
0
https://stackoverflow.com/questions/67989386/segmentation-fault-using-devtoolset-10-with-boost-1-73-static-libraries-complied
62,573,093
ambari on HDP cluster + ambari-metrics-collector service not start
we have some issue with ambari-metrics-collector service , ( we have HDP cluster version - 2.6.4 with 8 nodes ) ambari metrics collector service can’t start or start of few second then failed the details about metrics collector version rpm -qa | grep metrics ambari-metrics-grafana-2.6.1.0-143.x86_64 ambari-metrics-monitor-2.6.1.0-143.x86_64 ambari-metrics-collector-2.5.0.3-7.x86_64 ambari-metrics-hadoop-sink-2.6.1.0-143.x86_64 all machines are rhel 7.2 we performed the following steps in order to resolve the problem 1.restart metrics-collector service su - ams -c '/usr/sbin/ambari-metrics-collector --config /etc/ambari-metrics-collector/conf/ stop' su - ams -c '/usr/sbin/ambari-metrics-collector --config /etc/ambari-metrics-collector/conf/ start' or ambari-metrics-collector stop ambari-metrics-collector start 2.restart ambari-metrics-monitor on all nodes ambari-metrics-monitor stop ambari-metrics-monitor start 3.clean the folder /var/lib/ambari-metrics-collector/hbase-tmp/zookeeper/ mv /var/lib/ambari-metrics-collector/hbase-tmp/zookeeper/zookeeper_0 /tmp/bck/zookeeper/ Then restart metrics-collector service 4.Tuning the metrics-collector parameters according - [URL] we update the follwing parameters in ambari metrics_collector_heap_size=1024 hbase_regionserver_heapsize=1024 hbase_master_heapsize=512 hbase_master_xmn_size=128 status for now: - steps 1-4 doesn’t help From the logs we can see the following: log file - ambari-metrics-collector.log 2020-06-25 09:06:14,474 WARN org.apache.zookeeper.ClientCnxn: Session 0x172eab71f310002 for server null, unexpected error, closing socket connection and attempting reconnect java.net.ConnectException: Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:361) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1141) 2020-06-25 09:06:14,575 WARN org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Possibly transient ZooKeeper, quorum=master02.sys671.com:61181, exception=org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /ams-hbase-unsecure/meta-region-server log file - hbase-ams-master-master02.sys671.com.log 2020-06-25 09:38:18,799 WARN [RS:0;master02:51842-SendThread(master02.sys671.com:61181)] zookeeper.ClientCnxn: Session 0x172ead5d73a0004 for server null, unexpected error, closing socket connection and attempting reconnect java.net.ConnectException: Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:361) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1125) 2020-06-25 09:38:20,437 INFO [main-SendThread(master02.sys671.com:61181)] zookeeper.ClientCnxn: Opening socket connection to server master02.sys671.com/23.2.35.171:61181. Will not attempt to authenticate using SASL (unknown error) 2020-06-25 09:38:20,438 WARN [main-SendThread(master02.sys671.com:61181)] zookeeper.ClientCnxn: Session 0x172ead5d73a0002 for server null, unexpected error, closing socket connection and attempting reconnect java.net.ConnectException: Connection refused we also not see that port is listening ( timeline.metrics.service.webapp.address ) netstat -tulpn | grep 6188 any advice how to continue from this point ? we'll appreciate to get any help about this problem
ambari on HDP cluster + ambari-metrics-collector service not start we have some issue with ambari-metrics-collector service , ( we have HDP cluster version - 2.6.4 with 8 nodes ) ambari metrics collector service can’t start or start of few second then failed the details about metrics collector version rpm -qa | grep metrics ambari-metrics-grafana-2.6.1.0-143.x86_64 ambari-metrics-monitor-2.6.1.0-143.x86_64 ambari-metrics-collector-2.5.0.3-7.x86_64 ambari-metrics-hadoop-sink-2.6.1.0-143.x86_64 all machines are rhel 7.2 we performed the following steps in order to resolve the problem 1.restart metrics-collector service su - ams -c '/usr/sbin/ambari-metrics-collector --config /etc/ambari-metrics-collector/conf/ stop' su - ams -c '/usr/sbin/ambari-metrics-collector --config /etc/ambari-metrics-collector/conf/ start' or ambari-metrics-collector stop ambari-metrics-collector start 2.restart ambari-metrics-monitor on all nodes ambari-metrics-monitor stop ambari-metrics-monitor start 3.clean the folder /var/lib/ambari-metrics-collector/hbase-tmp/zookeeper/ mv /var/lib/ambari-metrics-collector/hbase-tmp/zookeeper/zookeeper_0 /tmp/bck/zookeeper/ Then restart metrics-collector service 4.Tuning the metrics-collector parameters according - [URL] we update the follwing parameters in ambari metrics_collector_heap_size=1024 hbase_regionserver_heapsize=1024 hbase_master_heapsize=512 hbase_master_xmn_size=128 status for now: - steps 1-4 doesn’t help From the logs we can see the following: log file - ambari-metrics-collector.log 2020-06-25 09:06:14,474 WARN org.apache.zookeeper.ClientCnxn: Session 0x172eab71f310002 for server null, unexpected error, closing socket connection and attempting reconnect java.net.ConnectException: Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:361) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1141) 2020-06-25 09:06:14,575 WARN org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Possibly transient ZooKeeper, quorum=master02.sys671.com:61181, exception=org.apache.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /ams-hbase-unsecure/meta-region-server log file - hbase-ams-master-master02.sys671.com.log 2020-06-25 09:38:18,799 WARN [RS:0;master02:51842-SendThread(master02.sys671.com:61181)] zookeeper.ClientCnxn: Session 0x172ead5d73a0004 for server null, unexpected error, closing socket connection and attempting reconnect java.net.ConnectException: Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717) at org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:361) at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1125) 2020-06-25 09:38:20,437 INFO [main-SendThread(master02.sys671.com:61181)] zookeeper.ClientCnxn: Opening socket connection to server master02.sys671.com/23.2.35.171:61181. Will not attempt to authenticate using SASL (unknown error) 2020-06-25 09:38:20,438 WARN [main-SendThread(master02.sys671.com:61181)] zookeeper.ClientCnxn: Session 0x172ead5d73a0002 for server null, unexpected error, closing socket connection and attempting reconnect java.net.ConnectException: Connection refused we also not see that port is listening ( timeline.metrics.service.webapp.address ) netstat -tulpn | grep 6188 any advice how to continue from this point ? we'll appreciate to get any help about this problem
metrics, rhel, ambari, hdp
3
2,254
0
https://stackoverflow.com/questions/62573093/ambari-on-hdp-cluster-ambari-metrics-collector-service-not-start
62,518,962
pip + why pip removing the current version while we not identify the option force-reinstall
I am installing the following python module pip install ipaddress-1.0.23-py2.py3-none-any.whl the pip failed on Processing ipaddress-1.0.23-py2.py3-none-any.whl Installing collected packages: ipaddress Attempting uninstall: ipaddress Found existing installation: ipaddress 1.0.16 ERROR: Cannot uninstall 'ipaddress'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall. What isn’t clear here why pip trying to uninstall the current pkg - ipaddress 1.0.16 As all can see we not using the option --force-reinstall So why pip install is removing the current - ipaddress 1.0.16 ? Reference - Difference between pip install options "ignore-installed" and "force-reinstall"
pip + why pip removing the current version while we not identify the option force-reinstall I am installing the following python module pip install ipaddress-1.0.23-py2.py3-none-any.whl the pip failed on Processing ipaddress-1.0.23-py2.py3-none-any.whl Installing collected packages: ipaddress Attempting uninstall: ipaddress Found existing installation: ipaddress 1.0.16 ERROR: Cannot uninstall 'ipaddress'. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall. What isn’t clear here why pip trying to uninstall the current pkg - ipaddress 1.0.16 As all can see we not using the option --force-reinstall So why pip install is removing the current - ipaddress 1.0.16 ? Reference - Difference between pip install options "ignore-installed" and "force-reinstall"
python, python-2.7, pip, rhel
3
2,141
2
https://stackoverflow.com/questions/62518962/pip-why-pip-removing-the-current-version-while-we-not-identify-the-option-forc
55,005,955
error: command &#39;gcc&#39; failed with exit status 1 while installing GDAL for Python3 on AWS EC2
I'm trying to install GDAL to use with Python 3 on an AWS EC2 Amazon2 machine (RHEL7). Following the instructions on this page: Install GDAL on Python 2.7 on a Amazon Linux virtual server I managed to get it working for python 2.7 ("import gdal from osgeo" works). However, I need it to work for python3 installed on my machine. I tried to adjust the instructions accordingly, and got to this set of commands: sudo yum-config-manager --enable epel sudo yum -y install make automake gcc gcc-c++ libcurl-devel proj-devel geos-devel gdal-devel gdal-python python3-devel sudo yum -y update cd /tmp curl -L [URL] | tar zxf - cd gdal-2.4.0/ ./configure --prefix=/usr/local --with-python sudo make -j4 sudo make install python3 -m pip install numpy --user sudo -H pip3 install --upgrade pip CPLUS_INCLUDE_PATH=/usr/include/gdal C_INCLUDE_PATH=/usr/include/gdal sudo python3 -m easy_install GDAL -d /usr/lib/python3.7/site-packages --include-dirs=/usr/include/gdal when I run the last command, the installation tries to run but fails with: "error: command 'gcc' failed with exit status 1". There are many errors before that, such as: /usr/local/include/cpl_vsi_error.h:58:24: error: ‘CPL_FORMAT_STRING’ has not been declared CPL_FORMAT_STRING(const char *fmt), ... ) ^~~~~~~~~~~~~~~~~ In file included from extensions/gdal_wrap.cpp:3173:0: From what I gather, the gcc can't find the GDAL headers, but it is specified in the command and I also did the including commands above: CPLUS_INCLUDE_PATH=/usr/include/gdal C_INCLUDE_PATH=/usr/include/gdal Any ideas will be helpful!
error: command &#39;gcc&#39; failed with exit status 1 while installing GDAL for Python3 on AWS EC2 I'm trying to install GDAL to use with Python 3 on an AWS EC2 Amazon2 machine (RHEL7). Following the instructions on this page: Install GDAL on Python 2.7 on a Amazon Linux virtual server I managed to get it working for python 2.7 ("import gdal from osgeo" works). However, I need it to work for python3 installed on my machine. I tried to adjust the instructions accordingly, and got to this set of commands: sudo yum-config-manager --enable epel sudo yum -y install make automake gcc gcc-c++ libcurl-devel proj-devel geos-devel gdal-devel gdal-python python3-devel sudo yum -y update cd /tmp curl -L [URL] | tar zxf - cd gdal-2.4.0/ ./configure --prefix=/usr/local --with-python sudo make -j4 sudo make install python3 -m pip install numpy --user sudo -H pip3 install --upgrade pip CPLUS_INCLUDE_PATH=/usr/include/gdal C_INCLUDE_PATH=/usr/include/gdal sudo python3 -m easy_install GDAL -d /usr/lib/python3.7/site-packages --include-dirs=/usr/include/gdal when I run the last command, the installation tries to run but fails with: "error: command 'gcc' failed with exit status 1". There are many errors before that, such as: /usr/local/include/cpl_vsi_error.h:58:24: error: ‘CPL_FORMAT_STRING’ has not been declared CPL_FORMAT_STRING(const char *fmt), ... ) ^~~~~~~~~~~~~~~~~ In file included from extensions/gdal_wrap.cpp:3173:0: From what I gather, the gcc can't find the GDAL headers, but it is specified in the command and I also did the including commands above: CPLUS_INCLUDE_PATH=/usr/include/gdal C_INCLUDE_PATH=/usr/include/gdal Any ideas will be helpful!
python-3.x, gcc, amazon-ec2, gdal, rhel
3
692
0
https://stackoverflow.com/questions/55005955/error-command-gcc-failed-with-exit-status-1-while-installing-gdal-for-python3
53,178,741
Chrony lose connection with local host after 10 minutes
I'm trying to configure chrony on RHEL 7.4. I configured the local clock (127.127.1.0) as a source, and chrony get time from it for about 10 minutes and then chrony stops receiving packets from it ('chronyc sources' shows 0 packets reached), although using tcpdump shows that ntp packets are sent to the local clock and back to localhost (also running 'chronyc ntpdata' shows it does not get time from the local clock). My chrony configuration: server 127.127.1.0 iburst peer 192.168.11.7 iburst local stratum 10 driftfile /var/lib/chrony/drift makestep 1.0 3 rtcsync 'chronyc sources' while it works: After 20 minutes: Also using tcp dump I figured that after starting chrony the packets from 127.127.1.0:123 looks correct, stratum and reference-id looks good, and after a few minutes the stratum written in the packets grow up to 15 and then all the later packets have "Stratum 0 (unspecified)" and also "Reference-ID: (unspec)". Any ideas?
Chrony lose connection with local host after 10 minutes I'm trying to configure chrony on RHEL 7.4. I configured the local clock (127.127.1.0) as a source, and chrony get time from it for about 10 minutes and then chrony stops receiving packets from it ('chronyc sources' shows 0 packets reached), although using tcpdump shows that ntp packets are sent to the local clock and back to localhost (also running 'chronyc ntpdata' shows it does not get time from the local clock). My chrony configuration: server 127.127.1.0 iburst peer 192.168.11.7 iburst local stratum 10 driftfile /var/lib/chrony/drift makestep 1.0 3 rtcsync 'chronyc sources' while it works: After 20 minutes: Also using tcp dump I figured that after starting chrony the packets from 127.127.1.0:123 looks correct, stratum and reference-id looks good, and after a few minutes the stratum written in the packets grow up to 15 and then all the later packets have "Stratum 0 (unspecified)" and also "Reference-ID: (unspec)". Any ideas?
rhel, ntp
3
956
0
https://stackoverflow.com/questions/53178741/chrony-lose-connection-with-local-host-after-10-minutes
51,701,969
Move a Perl script with dependencies from one computer to another
I have a Perl script on an old RHEL4 x86_64 Linux computer. This script runs only on this computer, and we'd like to run it on other Linux computers. The issue is there are a ton of Perl .pm and .xs files needed. I've done the following Created a virtual machine that has the same directory structure as the RHEL4 x86_64 Linux computer. I have run into errors this way, namely the Config.pm file aborting. Used the PAR Packager pp command to package the Perl script. I keep running into errors when packing it. What other routes can I take in executing the script on other Linux computers?
Move a Perl script with dependencies from one computer to another I have a Perl script on an old RHEL4 x86_64 Linux computer. This script runs only on this computer, and we'd like to run it on other Linux computers. The issue is there are a ton of Perl .pm and .xs files needed. I've done the following Created a virtual machine that has the same directory structure as the RHEL4 x86_64 Linux computer. I have run into errors this way, namely the Config.pm file aborting. Used the PAR Packager pp command to package the Perl script. I keep running into errors when packing it. What other routes can I take in executing the script on other Linux computers?
linux, perl, rhel
3
267
0
https://stackoverflow.com/questions/51701969/move-a-perl-script-with-dependencies-from-one-computer-to-another
51,637,889
How to install NodeJS on RHEL 7.5 server
How to install NodeJS on RHEL server 7.5? I have root privileges and tried first: curl -sL [URL] | bash - then: yum install -y nodejs Got error: Loaded plugins: amazon-id, product-id, rhui-lb, search-disabled-repos, subscription-manager This system is not registered with an entitlement server. You can use subscription-manager to register. [URL] [Errno 14] curl#22 - "The requested URL returned error: 407" Trying other mirror. No package nodejs available. Error: Nothing to do What is recommended way to install NodeJS on RHEL/CentOS servers and how to solve the problem above? Thanks.
How to install NodeJS on RHEL 7.5 server How to install NodeJS on RHEL server 7.5? I have root privileges and tried first: curl -sL [URL] | bash - then: yum install -y nodejs Got error: Loaded plugins: amazon-id, product-id, rhui-lb, search-disabled-repos, subscription-manager This system is not registered with an entitlement server. You can use subscription-manager to register. [URL] [Errno 14] curl#22 - "The requested URL returned error: 407" Trying other mirror. No package nodejs available. Error: Nothing to do What is recommended way to install NodeJS on RHEL/CentOS servers and how to solve the problem above? Thanks.
node.js, linux, centos, yum, rhel
3
2,661
0
https://stackoverflow.com/questions/51637889/how-to-install-nodejs-on-rhel-7-5-server
50,533,975
AWS EC2 RHEL upgrade from 6 to 7
Is it possible to upgrade from RHEL 6 to RHEL 7 on an EC2 instance ? The standard upgrade path suggested by rhel here requires that the machine is registered to receive updates from Subscription Management. And the subscription status on my machine says Unknown . # subscription-manager list +-------------------------------------------+ Installed Product Status +-------------------------------------------+ Product Name: Red Hat Enterprise Linux Server Product ID: 69 Version: 6.8 Beta Arch: x86_64 Status: Unknown Status Details: Starts: Ends: Has anybody been able to successfully upgrade it on EC2 instances?
AWS EC2 RHEL upgrade from 6 to 7 Is it possible to upgrade from RHEL 6 to RHEL 7 on an EC2 instance ? The standard upgrade path suggested by rhel here requires that the machine is registered to receive updates from Subscription Management. And the subscription status on my machine says Unknown . # subscription-manager list +-------------------------------------------+ Installed Product Status +-------------------------------------------+ Product Name: Red Hat Enterprise Linux Server Product ID: 69 Version: 6.8 Beta Arch: x86_64 Status: Unknown Status Details: Starts: Ends: Has anybody been able to successfully upgrade it on EC2 instances?
amazon-web-services, amazon-ec2, upgrade, rhel, rhel6
3
2,378
1
https://stackoverflow.com/questions/50533975/aws-ec2-rhel-upgrade-from-6-to-7
30,351,878
How to Install or enable PHP-FTP support on Linux Apache server?
In my PHP code i am using the below function ftp_connect() but it throws an error, FTP is not enabled for PHP. can any one tell me how to enable/install FTP in RHEL, Apache2.2 server,PHP Version is 5.3.3. This server is Production now my concern is if i reinstall PHP is it going to disturb my current environment?
How to Install or enable PHP-FTP support on Linux Apache server? In my PHP code i am using the below function ftp_connect() but it throws an error, FTP is not enabled for PHP. can any one tell me how to enable/install FTP in RHEL, Apache2.2 server,PHP Version is 5.3.3. This server is Production now my concern is if i reinstall PHP is it going to disturb my current environment?
php, tomcat, ftp, rhel
3
2,666
0
https://stackoverflow.com/questions/30351878/how-to-install-or-enable-php-ftp-support-on-linux-apache-server
29,681,618
installing rpm packages inside of a virtual python environment (virtualenv or pyvenv) on rhel 6 with no internet access
My RHEL server only has access to packages that are pre-approved by my IT. I'm developing an Python application that will be hosted on this server. For now, I've managed to build the latest version of python at a different location (/usr/local/bin). I'm able create virtual environments to isolate it from the rest of the system. Now, I need to install packages using pip in the environment to get the app to work eg. Django. Since every access to the internet has been disabled by IT, I cannot do pip install package_name. Well, I can but it tries to download the package from the internet and times out. When I download the rpm of the django package and try to install it in the virtual environment, it fails: (envtest) [root@arlwebimgakp02 envtest]# rpm -ivh python3-django-1.7.2-1.fc22.noarch.rpm error: Failed dependencies: /usr/bin/python3 is needed by python3-django-1.7.2-1.fc22.noarch python(abi) = 3.4 is needed by python3-django-1.7.2-1.fc22.noarch python-django-bash-completion = 1.7.2-1.fc22 is needed by python3-django-1.7.2-1.fc22.noarch python3 is needed by python3-django-1.7.2-1.fc22.noarch So it looks like the rpm install is looking for dependencies from the universal install of python (2.6.6) instead of the separately compiled 3.4.3 which is used in the virtual environment. Can I even do this? Can I install an rpm package to just a virtual environment? How else can I get my python packages to install in my virtual environments without internet access to the server?
installing rpm packages inside of a virtual python environment (virtualenv or pyvenv) on rhel 6 with no internet access My RHEL server only has access to packages that are pre-approved by my IT. I'm developing an Python application that will be hosted on this server. For now, I've managed to build the latest version of python at a different location (/usr/local/bin). I'm able create virtual environments to isolate it from the rest of the system. Now, I need to install packages using pip in the environment to get the app to work eg. Django. Since every access to the internet has been disabled by IT, I cannot do pip install package_name. Well, I can but it tries to download the package from the internet and times out. When I download the rpm of the django package and try to install it in the virtual environment, it fails: (envtest) [root@arlwebimgakp02 envtest]# rpm -ivh python3-django-1.7.2-1.fc22.noarch.rpm error: Failed dependencies: /usr/bin/python3 is needed by python3-django-1.7.2-1.fc22.noarch python(abi) = 3.4 is needed by python3-django-1.7.2-1.fc22.noarch python-django-bash-completion = 1.7.2-1.fc22 is needed by python3-django-1.7.2-1.fc22.noarch python3 is needed by python3-django-1.7.2-1.fc22.noarch So it looks like the rpm install is looking for dependencies from the universal install of python (2.6.6) instead of the separately compiled 3.4.3 which is used in the virtual environment. Can I even do this? Can I install an rpm package to just a virtual environment? How else can I get my python packages to install in my virtual environments without internet access to the server?
python, django, pip, rhel
3
5,258
3
https://stackoverflow.com/questions/29681618/installing-rpm-packages-inside-of-a-virtual-python-environment-virtualenv-or-py
20,316,362
no such repository on migrating to a new cvs server
I am moving from cvsserv1 to cvsserv2. I am running cvs1.11 on current server on RHEL. I am moving to cvsserv2 which is running ubuntu 12. This is my procedure to port cvs: zip entire repository on cvsserv1 move zip to cvsserv2 extract zip to /home/users on cvsserv2. setup cvs service on cvsserve2 in pserver mode. initialize repository on /home/users/cvsroot by using "cvs -d /home/users/cvsroot init" connect to cvsserv2 from eclipse using anonymous access to do a test checkout. I am failing on step6 with the error message "no such repository". What am I doing wrong? UPDATE I tried to change the above method, by adopting this [URL] and I was partially successful. At step 3 (as in that link), after initializing repo on cvsserv2, I copied my repository to /var/lib/cvsd/project1, overwriting CVSROOT folder. Now, after finishing all steps, I was able to connect successfully. However, when I try to check out, I don't see any branches. When I tried to Refresh Tags, I receive the following error: What is going wrong?
no such repository on migrating to a new cvs server I am moving from cvsserv1 to cvsserv2. I am running cvs1.11 on current server on RHEL. I am moving to cvsserv2 which is running ubuntu 12. This is my procedure to port cvs: zip entire repository on cvsserv1 move zip to cvsserv2 extract zip to /home/users on cvsserv2. setup cvs service on cvsserve2 in pserver mode. initialize repository on /home/users/cvsroot by using "cvs -d /home/users/cvsroot init" connect to cvsserv2 from eclipse using anonymous access to do a test checkout. I am failing on step6 with the error message "no such repository". What am I doing wrong? UPDATE I tried to change the above method, by adopting this [URL] and I was partially successful. At step 3 (as in that link), after initializing repo on cvsserv2, I copied my repository to /var/lib/cvsd/project1, overwriting CVSROOT folder. Now, after finishing all steps, I was able to connect successfully. However, when I try to check out, I don't see any branches. When I tried to Refresh Tags, I receive the following error: What is going wrong?
linux, ubuntu, migration, cvs, rhel
3
1,826
1
https://stackoverflow.com/questions/20316362/no-such-repository-on-migrating-to-a-new-cvs-server
19,842,110
yum update from wrong repository i
I am running a AWS Linux AMI in AWS, it seems to be RHEL under the covers and uses yum (kernel 3.4.62-53.42.amzn1.x86_64) . The amazon repo includes nginx 1.4.2 which is out of date so I removed it and re-installed using the repo given at nginx website. I did this as such: sudo yum remove nginx* sudo yum --disablerepo="*" --enablerepo="nginx" install nginx Note that yum priorities are disabled. I tried messing around with setting priority= in the repo files but decided disabling priorities made sense after reading up on it. /etc/yum.repos.d files : amzn-main.repo contains: [amzn-main] name=amzn-main-Base mirrorlist=[URL] mirror_expire=300 metadata_expire=300 priority=10 failovermethod=priority fastestmirror_enabled=0 gpgcheck=1 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-amazon-ga enabled=1 retries=5 timeout=10 report_instanceid=yes [amzn-main-debuginfo] name=amzn-main-debuginfo mirrorlist=[URL] mirror_expire=300 metadata_expire=300 priority=10 failovermethod=priority fastestmirror_enabled=0 gpgcheck=1 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-amazon-ga enabled=0 retries=5 timeout=10 report_instanceid=yes nginx.repo contains: [nginx] name=nginx repo baseurl=[URL] gpgcheck=0 enabled=1 priority=1 Now doing the update once nginx 1.4.3 is installed, it tries to install the old version from the amazon main repo: $ sudo yum update Loaded plugins: update-motd, upgrade-helper Resolving Dependencies --> Running transaction check ---> Package nginx.x86_64 0:1.4.3-1.el6.ngx will be updated ---> Package nginx.x86_64 1:1.4.2-1.12.amzn1 will be an update --> Finished Dependency Resolution Dependencies Resolved So not sure why its picking up the older version. Is there a way to configure yum to only update if the version is newer than currently installed? If there is no way, is there a way to skip nginx package when I do a general update and then update nginx using another command as described above that disables all repos and uses only nginx repo, here it is again: sudo yum --disablerepo="*" --enablerepo="nginx" install nginx
yum update from wrong repository i I am running a AWS Linux AMI in AWS, it seems to be RHEL under the covers and uses yum (kernel 3.4.62-53.42.amzn1.x86_64) . The amazon repo includes nginx 1.4.2 which is out of date so I removed it and re-installed using the repo given at nginx website. I did this as such: sudo yum remove nginx* sudo yum --disablerepo="*" --enablerepo="nginx" install nginx Note that yum priorities are disabled. I tried messing around with setting priority= in the repo files but decided disabling priorities made sense after reading up on it. /etc/yum.repos.d files : amzn-main.repo contains: [amzn-main] name=amzn-main-Base mirrorlist=[URL] mirror_expire=300 metadata_expire=300 priority=10 failovermethod=priority fastestmirror_enabled=0 gpgcheck=1 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-amazon-ga enabled=1 retries=5 timeout=10 report_instanceid=yes [amzn-main-debuginfo] name=amzn-main-debuginfo mirrorlist=[URL] mirror_expire=300 metadata_expire=300 priority=10 failovermethod=priority fastestmirror_enabled=0 gpgcheck=1 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-amazon-ga enabled=0 retries=5 timeout=10 report_instanceid=yes nginx.repo contains: [nginx] name=nginx repo baseurl=[URL] gpgcheck=0 enabled=1 priority=1 Now doing the update once nginx 1.4.3 is installed, it tries to install the old version from the amazon main repo: $ sudo yum update Loaded plugins: update-motd, upgrade-helper Resolving Dependencies --> Running transaction check ---> Package nginx.x86_64 0:1.4.3-1.el6.ngx will be updated ---> Package nginx.x86_64 1:1.4.2-1.12.amzn1 will be an update --> Finished Dependency Resolution Dependencies Resolved So not sure why its picking up the older version. Is there a way to configure yum to only update if the version is newer than currently installed? If there is no way, is there a way to skip nginx package when I do a general update and then update nginx using another command as described above that disables all repos and uses only nginx repo, here it is again: sudo yum --disablerepo="*" --enablerepo="nginx" install nginx
amazon-web-services, nginx, yum, rhel
3
3,433
1
https://stackoverflow.com/questions/19842110/yum-update-from-wrong-repository-i
18,307,208
Incomplete output with psycopg2 and csv writer in python
I wrote the following python code that would write a portion of a postgresql table to a csv file using psycopg2 and python's csvwriter. The problem that I am having is that I get an incomplete csv file and I am not sure why. Typically, what I see is about 80 to 90% of the table in the csv file and usually the last row outputted in the csv file is incomplete as well. Running Python2.6.6 on RHEL6.4. Looking for some help for a work around on this issue. import psycopg2 import csv connection = psycopg2.connect('dbname=db user=user password=pass host=localhost port=5555') cursor = connection.cursor() cursor.execute('SELECT title,date,publisher,author FROM books;') with open('books.csv', 'wb') as csvfile: csvwriter = csv.writer(csvfile,delimiter=',',quotechar='|', quoting=csv.QUOTE_MINIMAL) csvwriter.writerow(['title,date,publisher,author']) for record in cursor: csvwriter.writerow(record) cursor.close() connection.close()
Incomplete output with psycopg2 and csv writer in python I wrote the following python code that would write a portion of a postgresql table to a csv file using psycopg2 and python's csvwriter. The problem that I am having is that I get an incomplete csv file and I am not sure why. Typically, what I see is about 80 to 90% of the table in the csv file and usually the last row outputted in the csv file is incomplete as well. Running Python2.6.6 on RHEL6.4. Looking for some help for a work around on this issue. import psycopg2 import csv connection = psycopg2.connect('dbname=db user=user password=pass host=localhost port=5555') cursor = connection.cursor() cursor.execute('SELECT title,date,publisher,author FROM books;') with open('books.csv', 'wb') as csvfile: csvwriter = csv.writer(csvfile,delimiter=',',quotechar='|', quoting=csv.QUOTE_MINIMAL) csvwriter.writerow(['title,date,publisher,author']) for record in cursor: csvwriter.writerow(record) cursor.close() connection.close()
python, csv, python-2.6, psycopg2, rhel
3
1,018
0
https://stackoverflow.com/questions/18307208/incomplete-output-with-psycopg2-and-csv-writer-in-python
17,527,377
Using virt-install to setup an Ubuntu 12.04 VM
I am trying to set up an Ubuntu precise image with this command: virt-install --virt-type kvm --name precise --ram 1024 --cdrom=/data/isos/precise-64-mini.iso --disk /tmp/precise.qcow2,format=qcow2 --network network=default --graphics vnc,listen=0.0.0.0 --noautoconsole --os-type=linux --os-variant=ubuntuprecise But I get this error every time: ERROR OS variant 'ubuntuprecise' does not exist in our dictionary for OS type 'linux' And I noticed that 'ubuntuprecise' is not listed in the variant lists. Is there any way to add/update the list or is there an alternative to use?
Using virt-install to setup an Ubuntu 12.04 VM I am trying to set up an Ubuntu precise image with this command: virt-install --virt-type kvm --name precise --ram 1024 --cdrom=/data/isos/precise-64-mini.iso --disk /tmp/precise.qcow2,format=qcow2 --network network=default --graphics vnc,listen=0.0.0.0 --noautoconsole --os-type=linux --os-variant=ubuntuprecise But I get this error every time: ERROR OS variant 'ubuntuprecise' does not exist in our dictionary for OS type 'linux' And I noticed that 'ubuntuprecise' is not listed in the variant lists. Is there any way to add/update the list or is there an alternative to use?
virtual-machine, ubuntu-12.04, rhel, kvm
3
766
0
https://stackoverflow.com/questions/17527377/using-virt-install-to-setup-an-ubuntu-12-04-vm
10,797,413
Failed Kerberos authentication when Debug logging enabled
I am currently running a RHEL 6 server with IPA and a user test_user defined, attempting to authenticate using the 'Krb5LoginModule' Login Module I am getting the exception at the bottom of this question. HOWEVER - if I add the system property -Dsun.security.krb5.debug=true the authentication completes without error - that system property is the only change between the failure scenario and successful scenario. The same username and password also work flawlessly in kinit from the same machine running the Java client. This is using Oracle jdk1.6.0_32. So the question is what else is different when debug logging is enabled that would change the behaviour of the login process? At the moment enabling debug logging to understand the failure didn't work as that in itself caused it to work. Exception in thread "main" javax.security.auth.login.LoginException: Integrity check on decrypted field failed (31) - PREAUTH_FAILED at com.sun.security.auth.module.Krb5LoginModule.attemptAuthentication(Krb5LoginModule.java:696) at com.sun.security.auth.module.Krb5LoginModule.login(Krb5LoginModule.java:542) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at javax.security.auth.login.LoginContext.invoke(LoginContext.java:769) at javax.security.auth.login.LoginContext.access$000(LoginContext.java:186) at javax.security.auth.login.LoginContext$5.run(LoginContext.java:706) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.login.LoginContext.invokeCreatorPriv(LoginContext.java:703) at javax.security.auth.login.LoginContext.login(LoginContext.java:575) at com.darranl.as.sasl.gssapi.KerberosLoginUtil.login(KerberosLoginUtil.java:50) at com.darranl.as.sasl.gssapi.KerberosLoginUtil.main(KerberosLoginUtil.java:131) Caused by: KrbException: Integrity check on decrypted field failed (31) - PREAUTH_FAILED at sun.security.krb5.KrbAsRep.<init>(KrbAsRep.java:72) at sun.security.krb5.KrbAsReq.getReply(KrbAsReq.java:446) at sun.security.krb5.Credentials.sendASRequest(Credentials.java:401) at sun.security.krb5.Credentials.acquireTGT(Credentials.java:373) at com.sun.security.auth.module.Krb5LoginModule.attemptAuthentication(Krb5LoginModule.java:662) ... 13 more Caused by: KrbException: Identifier doesn't match expected value (906) at sun.security.krb5.internal.KDCRep.init(KDCRep.java:133) at sun.security.krb5.internal.ASRep.init(ASRep.java:58) at sun.security.krb5.internal.ASRep.<init>(ASRep.java:53) at sun.security.krb5.KrbAsRep.<init>(KrbAsRep.java:50) ... 17 more
Failed Kerberos authentication when Debug logging enabled I am currently running a RHEL 6 server with IPA and a user test_user defined, attempting to authenticate using the 'Krb5LoginModule' Login Module I am getting the exception at the bottom of this question. HOWEVER - if I add the system property -Dsun.security.krb5.debug=true the authentication completes without error - that system property is the only change between the failure scenario and successful scenario. The same username and password also work flawlessly in kinit from the same machine running the Java client. This is using Oracle jdk1.6.0_32. So the question is what else is different when debug logging is enabled that would change the behaviour of the login process? At the moment enabling debug logging to understand the failure didn't work as that in itself caused it to work. Exception in thread "main" javax.security.auth.login.LoginException: Integrity check on decrypted field failed (31) - PREAUTH_FAILED at com.sun.security.auth.module.Krb5LoginModule.attemptAuthentication(Krb5LoginModule.java:696) at com.sun.security.auth.module.Krb5LoginModule.login(Krb5LoginModule.java:542) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at javax.security.auth.login.LoginContext.invoke(LoginContext.java:769) at javax.security.auth.login.LoginContext.access$000(LoginContext.java:186) at javax.security.auth.login.LoginContext$5.run(LoginContext.java:706) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.login.LoginContext.invokeCreatorPriv(LoginContext.java:703) at javax.security.auth.login.LoginContext.login(LoginContext.java:575) at com.darranl.as.sasl.gssapi.KerberosLoginUtil.login(KerberosLoginUtil.java:50) at com.darranl.as.sasl.gssapi.KerberosLoginUtil.main(KerberosLoginUtil.java:131) Caused by: KrbException: Integrity check on decrypted field failed (31) - PREAUTH_FAILED at sun.security.krb5.KrbAsRep.<init>(KrbAsRep.java:72) at sun.security.krb5.KrbAsReq.getReply(KrbAsReq.java:446) at sun.security.krb5.Credentials.sendASRequest(Credentials.java:401) at sun.security.krb5.Credentials.acquireTGT(Credentials.java:373) at com.sun.security.auth.module.Krb5LoginModule.attemptAuthentication(Krb5LoginModule.java:662) ... 13 more Caused by: KrbException: Identifier doesn't match expected value (906) at sun.security.krb5.internal.KDCRep.init(KDCRep.java:133) at sun.security.krb5.internal.ASRep.init(ASRep.java:58) at sun.security.krb5.internal.ASRep.<init>(ASRep.java:53) at sun.security.krb5.KrbAsRep.<init>(KrbAsRep.java:50) ... 17 more
java, kerberos, ipa, rhel
3
4,907
0
https://stackoverflow.com/questions/10797413/failed-kerberos-authentication-when-debug-logging-enabled
62,996,781
Java printing thread gets stuck : WHY?
In my Java 8 application (RHEL 6.x, Wildfly 10.1.0.Final) the first time a user prints a document, application gets stuck while getting the list of printers from the system. Here is the stacktrace of the blocking thread : "Thread-211" #799 daemon prio=5 os_prio=0 tid=0x00007fca543a6800 nid=0x10755 runnable [0x00007fca02820000] java.lang.Thread.State: RUNNABLE at sun.print.CUPSPrinter.canConnect(Native Method) at sun.print.CUPSPrinter.isCupsRunning(CUPSPrinter.java:444) at sun.print.UnixPrintServiceLookup.getDefaultPrintService(UnixPrintServiceLookup.java:650) - locked <0x00000006d2c7fff8> (a sun.print.UnixPrintServiceLookup) at sun.print.UnixPrintServiceLookup.refreshServices(UnixPrintServiceLookup.java:277) - locked <0x00000006d2c7fff8> (a sun.print.UnixPrintServiceLookup) at sun.print.UnixPrintServiceLookup$PrinterChangeListener.run(UnixPrintServiceLookup.java:947) Other users trying to print documents and relatives threads are blocked by this one. I looked at the source code of CUPSPrinter.canConnect() (native code) and at this point we try to connect to the cups server : /* * Checks if connection can be made to the server. * */ JNIEXPORT jboolean JNICALL Java_sun_print_CUPSPrinter_canConnect(JNIEnv *env, jobject printObj, jstring server, jint port) { const char *serverName; serverName = (*env)->GetStringUTFChars(env, server, NULL); if (serverName != NULL) { http_t *http = j2d_httpConnect(serverName, (int)port); (*env)->ReleaseStringUTFChars(env, server, serverName); if (http != NULL) { j2d_httpClose(http); return JNI_TRUE; } } return JNI_FALSE; } In my case CUPS is on the same host listening on port 631. I checked the logs & everything seems to be fine. I also checked active connections for cups with netstat : tcp 0 0 0.0.0.0:631 0.0.0.0:* LISTEN 76107/cupsd tcp 0 0 127.0.0.1:45652 127.0.0.1:631 TIME_WAIT - tcp 0 0 :::631 :::* LISTEN 76107/cupsd tcp 0 0 ::1:35982 ::1:631 TIME_WAIT - tcp 0 0 ::1:35981 ::1:631 TIME_WAIT - tcp 0 0 ::1:35978 ::1:631 TIME_WAIT - tcp 0 0 ::1:35979 ::1:631 TIME_WAIT - udp 0 0 0.0.0.0:631 0.0.0.0:* 76107/cupsd Important notes : If I restart Cups service, thread is not deblocked. It seems to live endessly until application restarts. I found a bug similar to this on Open JDK : [URL] But the workaround of setting -Dsun.java2d.print.polling=false does not work for me (the property seems to be cleared at some point for an obscure reason, so PrinterChangeListener gets instantiated and though polling is not desactivated). I can't reproduce the problem with a test application (clone of production) on the same server Please HELP !!
Java printing thread gets stuck : WHY? In my Java 8 application (RHEL 6.x, Wildfly 10.1.0.Final) the first time a user prints a document, application gets stuck while getting the list of printers from the system. Here is the stacktrace of the blocking thread : "Thread-211" #799 daemon prio=5 os_prio=0 tid=0x00007fca543a6800 nid=0x10755 runnable [0x00007fca02820000] java.lang.Thread.State: RUNNABLE at sun.print.CUPSPrinter.canConnect(Native Method) at sun.print.CUPSPrinter.isCupsRunning(CUPSPrinter.java:444) at sun.print.UnixPrintServiceLookup.getDefaultPrintService(UnixPrintServiceLookup.java:650) - locked <0x00000006d2c7fff8> (a sun.print.UnixPrintServiceLookup) at sun.print.UnixPrintServiceLookup.refreshServices(UnixPrintServiceLookup.java:277) - locked <0x00000006d2c7fff8> (a sun.print.UnixPrintServiceLookup) at sun.print.UnixPrintServiceLookup$PrinterChangeListener.run(UnixPrintServiceLookup.java:947) Other users trying to print documents and relatives threads are blocked by this one. I looked at the source code of CUPSPrinter.canConnect() (native code) and at this point we try to connect to the cups server : /* * Checks if connection can be made to the server. * */ JNIEXPORT jboolean JNICALL Java_sun_print_CUPSPrinter_canConnect(JNIEnv *env, jobject printObj, jstring server, jint port) { const char *serverName; serverName = (*env)->GetStringUTFChars(env, server, NULL); if (serverName != NULL) { http_t *http = j2d_httpConnect(serverName, (int)port); (*env)->ReleaseStringUTFChars(env, server, serverName); if (http != NULL) { j2d_httpClose(http); return JNI_TRUE; } } return JNI_FALSE; } In my case CUPS is on the same host listening on port 631. I checked the logs & everything seems to be fine. I also checked active connections for cups with netstat : tcp 0 0 0.0.0.0:631 0.0.0.0:* LISTEN 76107/cupsd tcp 0 0 127.0.0.1:45652 127.0.0.1:631 TIME_WAIT - tcp 0 0 :::631 :::* LISTEN 76107/cupsd tcp 0 0 ::1:35982 ::1:631 TIME_WAIT - tcp 0 0 ::1:35981 ::1:631 TIME_WAIT - tcp 0 0 ::1:35978 ::1:631 TIME_WAIT - tcp 0 0 ::1:35979 ::1:631 TIME_WAIT - udp 0 0 0.0.0.0:631 0.0.0.0:* 76107/cupsd Important notes : If I restart Cups service, thread is not deblocked. It seems to live endessly until application restarts. I found a bug similar to this on Open JDK : [URL] But the workaround of setting -Dsun.java2d.print.polling=false does not work for me (the property seems to be cleared at some point for an obscure reason, so PrinterChangeListener gets instantiated and though polling is not desactivated). I can't reproduce the problem with a test application (clone of production) on the same server Please HELP !!
java, multithreading, printing, rhel, cups
3
463
1
https://stackoverflow.com/questions/62996781/java-printing-thread-gets-stuck-why
60,460,705
Podman images not showing with podman image ls
I am trying to setup a build server in a Red Hat Enterprise Linux 8 (CentoOS 8) virtual machine. I installed podman by running sudo dnf install -y @container-tools I then ran sudo podman pull mcr.microsoft.com/dotnet/core/aspnet:3.1-buster-slim to pull a container image from docker: Trying to pull mcr.microsoft.com/dotnet/core/aspnet:3.1-buster-slim...Getting image source signatures Copying blob e936bd534ffb done Copying blob caf64655bcbb done Copying blob 4156e490f05f done Copying blob 68ced04f60ab done Copying blob 7064c3d93b4a done Copying config e2cd20adb1 done Writing manifest to image destination Storing signatures e2cd20adb1292ef24ca70de7abaddaadd57a5c932d3852b972e43b6f05a03dea This looks successful to me. And if I run it again, I get told that the layers "already exists". But then I run: podman image ls and I get an empty list back: REPOSITORY TAG IMAGE ID CREATED SIZE I also tried the following commands to get a list: podman image ls -a podman image list podman image list -a podman images podman images ls podman images ls -a podman images list podman images list -a They all give an empty list. How can I see the container image that I pulled down? Update: I ran sudo podman run --rm --name=linuxconfig-test -p 80:80 httpd and (on another machine) browsed to the ip address of my linux machine and got It Works! shown. So podman is working at least in part.
Podman images not showing with podman image ls I am trying to setup a build server in a Red Hat Enterprise Linux 8 (CentoOS 8) virtual machine. I installed podman by running sudo dnf install -y @container-tools I then ran sudo podman pull mcr.microsoft.com/dotnet/core/aspnet:3.1-buster-slim to pull a container image from docker: Trying to pull mcr.microsoft.com/dotnet/core/aspnet:3.1-buster-slim...Getting image source signatures Copying blob e936bd534ffb done Copying blob caf64655bcbb done Copying blob 4156e490f05f done Copying blob 68ced04f60ab done Copying blob 7064c3d93b4a done Copying config e2cd20adb1 done Writing manifest to image destination Storing signatures e2cd20adb1292ef24ca70de7abaddaadd57a5c932d3852b972e43b6f05a03dea This looks successful to me. And if I run it again, I get told that the layers "already exists". But then I run: podman image ls and I get an empty list back: REPOSITORY TAG IMAGE ID CREATED SIZE I also tried the following commands to get a list: podman image ls -a podman image list podman image list -a podman images podman images ls podman images ls -a podman images list podman images list -a They all give an empty list. How can I see the container image that I pulled down? Update: I ran sudo podman run --rm --name=linuxconfig-test -p 80:80 httpd and (on another machine) browsed to the ip address of my linux machine and got It Works! shown. So podman is working at least in part.
centos, containers, rhel, podman
2
23,792
4
https://stackoverflow.com/questions/60460705/podman-images-not-showing-with-podman-image-ls
50,354,012
systemctl failing with unknown section &#39;Timer&#39;
I have a systemctl job that performs vertica backup to s3, i wanted to add a timer that runs everyday at 3am . I tried to create an override file with the timer section but when i do daemon-reload, I am getting an error `Unknown section Timer', I am unable to find the issue. /etc/systemd/system/vertica-backup.service.d/Override.conf [Timer] OnCalendar=*-*-* 03:00:00 Unit=vertica-backup.service /etc/systemd/system/vertica-backup.service: [Unit] Description = Vertica Backup Service After = network.target [Service] User= dbadmin ExecStart= /usr/local/bin/vertica-backup.sh Error May 15 15:19:47 ip-10-150-4-42.ec2.internal systemd[1]: [/etc/systemd/system/vertica-backup.service.d/override.conf:1] Unknown section 'Timer'. Ignoring. May 15 15:19:50 ip-10-150-4-42.ec2.internal systemd[1]: [/etc/systemd/system/vertica-backup.service.d/override.conf:1] Unknown section 'Timer'. Ignoring.
systemctl failing with unknown section &#39;Timer&#39; I have a systemctl job that performs vertica backup to s3, i wanted to add a timer that runs everyday at 3am . I tried to create an override file with the timer section but when i do daemon-reload, I am getting an error `Unknown section Timer', I am unable to find the issue. /etc/systemd/system/vertica-backup.service.d/Override.conf [Timer] OnCalendar=*-*-* 03:00:00 Unit=vertica-backup.service /etc/systemd/system/vertica-backup.service: [Unit] Description = Vertica Backup Service After = network.target [Service] User= dbadmin ExecStart= /usr/local/bin/vertica-backup.sh Error May 15 15:19:47 ip-10-150-4-42.ec2.internal systemd[1]: [/etc/systemd/system/vertica-backup.service.d/override.conf:1] Unknown section 'Timer'. Ignoring. May 15 15:19:50 ip-10-150-4-42.ec2.internal systemd[1]: [/etc/systemd/system/vertica-backup.service.d/override.conf:1] Unknown section 'Timer'. Ignoring.
rhel, systemd, systemctl
2
6,278
2
https://stackoverflow.com/questions/50354012/systemctl-failing-with-unknown-section-timer
40,305,095
Cannot install rgdal package in R on RHEL6, unable to load shared object rgdal.so
I have AWS RHEL6 server where I've installed (compiled) rgdal 1.9.2 and proj packages. Now in R I want to install rgdal package but it gives this error: ** testing if installed package can be loaded Error in dyn.load(file, DLLpath = DLLpath, ...) : unable to load shared object '/home/Roel/R/x86_64-redhat-linux-gnu-library/3.3/rgdal/libs/rgdal.so': libgdal.so.1: cannot open shared object file: No such file or directory Error: loading failed Execution halted ERROR: loading failed * removing ‘/home/Roel/R/x86_64-redhat-linux-gnu-library/3.3/rgdal’ Warning in install.packages : installation of package ‘rgdal’ had non-zero exit status Full install log can be found here . When I do locate rgdal.so in Linux it doesn't find anything. I can also not find how to get rgdal.so of what it is part of. Has anyone any idea on how to fix this?
Cannot install rgdal package in R on RHEL6, unable to load shared object rgdal.so I have AWS RHEL6 server where I've installed (compiled) rgdal 1.9.2 and proj packages. Now in R I want to install rgdal package but it gives this error: ** testing if installed package can be loaded Error in dyn.load(file, DLLpath = DLLpath, ...) : unable to load shared object '/home/Roel/R/x86_64-redhat-linux-gnu-library/3.3/rgdal/libs/rgdal.so': libgdal.so.1: cannot open shared object file: No such file or directory Error: loading failed Execution halted ERROR: loading failed * removing ‘/home/Roel/R/x86_64-redhat-linux-gnu-library/3.3/rgdal’ Warning in install.packages : installation of package ‘rgdal’ had non-zero exit status Full install log can be found here . When I do locate rgdal.so in Linux it doesn't find anything. I can also not find how to get rgdal.so of what it is part of. Has anyone any idea on how to fix this?
r, linux, rhel, rhel6, rgdal
2
2,156
2
https://stackoverflow.com/questions/40305095/cannot-install-rgdal-package-in-r-on-rhel6-unable-to-load-shared-object-rgdal-s
16,817,925
Unicode character default collation table
I don't know which site this question belongs exactly, so posting it here. I use Postgresql 9.2 on RHEL 6.4 and observe the following: select foo from unnest('{а,ә,б,в,г,д,е,ж}'::text[]) as foo order by foo collate "kk_KZ.utf8" gives а ә б в г д е ж BUT select foo from unnest('{а,ә,б,в,г,д,е,ж}'::text[]) as foo order by foo collate "en_US.utf8" gives а б в г д е ә -- misplaced ж Further, I found that there is the Default Unicode Collation Element Table [1], which lists the character in question (04D9 ; [.199D.0020.0002.04D9] # CYRILLIC SMALL LETTER SCHWA) in proper order. I understand that it is silly to expect the cyrillic characters be handled properly by "en_US.utf8" locale, but what is the correct behavior by Unicode or any other relevant standards in cases, where a character does not normally belong to language/locale used for collation? [1] [URL]
Unicode character default collation table I don't know which site this question belongs exactly, so posting it here. I use Postgresql 9.2 on RHEL 6.4 and observe the following: select foo from unnest('{а,ә,б,в,г,д,е,ж}'::text[]) as foo order by foo collate "kk_KZ.utf8" gives а ә б в г д е ж BUT select foo from unnest('{а,ә,б,в,г,д,е,ж}'::text[]) as foo order by foo collate "en_US.utf8" gives а б в г д е ә -- misplaced ж Further, I found that there is the Default Unicode Collation Element Table [1], which lists the character in question (04D9 ; [.199D.0020.0002.04D9] # CYRILLIC SMALL LETTER SCHWA) in proper order. I understand that it is silly to expect the cyrillic characters be handled properly by "en_US.utf8" locale, but what is the correct behavior by Unicode or any other relevant standards in cases, where a character does not normally belong to language/locale used for collation? [1] [URL]
postgresql, unicode, collation, rhel
2
3,075
3
https://stackoverflow.com/questions/16817925/unicode-character-default-collation-table
43,947,124
postgresql createuser -interactive not working in CentOS 7
A fresh installation of CentOS 7 needs a fresh installation of PostgreSQL, with a new user and a new role. I am following the steps described in this tutorial to accomplish this goal. However, the terminal is not providing the interactive menu that the tutorial promises when I type createuser -interactive . Instead, I get the following blank prompt: [this_centos_user@localhost ~]$ sudo -i -u postgres [sudo] password for this_centos_user: -bash-4.2$ createuser –interactive -bash-4.2$ What specific commands need to be typed in order to get the interactive createuser interface to appear and let me give a username, password, etc.? The Specific Situation: 1.) First, I installed the postgresql-server package and the "contrib" package with the following command: sudo yum install postgresql-server postgresql-contrib 2.) Next, I created a new PostgreSQL database cluster: sudo postgresql-setup initdb 3.) I then set up password authentication editing PostgreSQL's host-based authentication (HBA) configuration by typing sudo vi /var/lib/pgsql/data/pg_hba.conf and changing the following lines to include md5 instead of ident: host all all 127.0.0.1/32 md5 host all all ::1/128 md5 4.) After saving and exiting vi , I started and enabled PostgreSQL with the following: sudo systemctl start postgresql sudo systemctl enable postgresql 5.) Next, I logged in to PostgreSQL with the postgres account that we created above, and tried to create the user with the code from the top of the OP above, as follows: [this_centos_user@localhost ~]$ sudo -i -u postgres [sudo] password for this_centos_user: -bash-4.2$ createuser –interactive -bash-4.2$ So how do I create this user?
postgresql createuser -interactive not working in CentOS 7 A fresh installation of CentOS 7 needs a fresh installation of PostgreSQL, with a new user and a new role. I am following the steps described in this tutorial to accomplish this goal. However, the terminal is not providing the interactive menu that the tutorial promises when I type createuser -interactive . Instead, I get the following blank prompt: [this_centos_user@localhost ~]$ sudo -i -u postgres [sudo] password for this_centos_user: -bash-4.2$ createuser –interactive -bash-4.2$ What specific commands need to be typed in order to get the interactive createuser interface to appear and let me give a username, password, etc.? The Specific Situation: 1.) First, I installed the postgresql-server package and the "contrib" package with the following command: sudo yum install postgresql-server postgresql-contrib 2.) Next, I created a new PostgreSQL database cluster: sudo postgresql-setup initdb 3.) I then set up password authentication editing PostgreSQL's host-based authentication (HBA) configuration by typing sudo vi /var/lib/pgsql/data/pg_hba.conf and changing the following lines to include md5 instead of ident: host all all 127.0.0.1/32 md5 host all all ::1/128 md5 4.) After saving and exiting vi , I started and enabled PostgreSQL with the following: sudo systemctl start postgresql sudo systemctl enable postgresql 5.) Next, I logged in to PostgreSQL with the postgres account that we created above, and tried to create the user with the code from the top of the OP above, as follows: [this_centos_user@localhost ~]$ sudo -i -u postgres [sudo] password for this_centos_user: -bash-4.2$ createuser –interactive -bash-4.2$ So how do I create this user?
linux, postgresql, centos, centos7, rhel
2
5,459
2
https://stackoverflow.com/questions/43947124/postgresql-createuser-interactive-not-working-in-centos-7
73,098,085
linux + how to measure ssh time until login to remote machine
we want to measure the time that is needed until ssh successfully login to remote machine is it possible to measure ssh time until login to remote machine? note - all machine are RHEL machines for example lets say we want to verify how much time is needed to login machine_X from machine_Y example from machine_Y while machine_X is machine_[1..100] ssh time until login to machine1 was 2 sec ssh time until login to machine2 was 6 sec ssh time until login to machine3 was 12 sec ssh time until login to machine4 was 3 sec . . .
linux + how to measure ssh time until login to remote machine we want to measure the time that is needed until ssh successfully login to remote machine is it possible to measure ssh time until login to remote machine? note - all machine are RHEL machines for example lets say we want to verify how much time is needed to login machine_X from machine_Y example from machine_Y while machine_X is machine_[1..100] ssh time until login to machine1 was 2 sec ssh time until login to machine2 was 6 sec ssh time until login to machine3 was 12 sec ssh time until login to machine4 was 3 sec . . .
bash, ssh, rhel
2
868
3
https://stackoverflow.com/questions/73098085/linux-how-to-measure-ssh-time-until-login-to-remote-machine
65,690,691
How to emulate class template argument deduction pre-C++17?
I am working on removing sscanf() calls from a C++ codebase and replacing them with the std::stringstream implementation described here: [URL] . The relevant code is: template<class Char> class imatch { const Char* s; public: imatch(const Char* x) :s(x) {} template<class Stream> friend Stream& operator >> (Stream& st, const imatch& m) { std::basic_string<Char> x; st >> x; //strip spaces, read chars up to space if(x!=m.s) st.setstate(st.failbit); //set as "failure" a mismatch return st; } }; Then in my codebase: std::stringstream ss("value = 15"); //the input int val=0; ss >> imatch("value") >> imatch("=") >> val; if(ss) { std::cout << "read value = " << val << std::endl; } else { std::cout << "read failed" << std::endl; } This uses class template argument deduction in the constructor calls. It works great...in C++17. The problem is that this code needs to compile all the way back to RHEL6, which only supports -std=c++0x at best (a subset of C++11). What is the most succinct way to write and use this class such that users can easily port their sscanf() calls to use it, without access to C++17?
How to emulate class template argument deduction pre-C++17? I am working on removing sscanf() calls from a C++ codebase and replacing them with the std::stringstream implementation described here: [URL] . The relevant code is: template<class Char> class imatch { const Char* s; public: imatch(const Char* x) :s(x) {} template<class Stream> friend Stream& operator >> (Stream& st, const imatch& m) { std::basic_string<Char> x; st >> x; //strip spaces, read chars up to space if(x!=m.s) st.setstate(st.failbit); //set as "failure" a mismatch return st; } }; Then in my codebase: std::stringstream ss("value = 15"); //the input int val=0; ss >> imatch("value") >> imatch("=") >> val; if(ss) { std::cout << "read value = " << val << std::endl; } else { std::cout << "read failed" << std::endl; } This uses class template argument deduction in the constructor calls. It works great...in C++17. The problem is that this code needs to compile all the way back to RHEL6, which only supports -std=c++0x at best (a subset of C++11). What is the most succinct way to write and use this class such that users can easily port their sscanf() calls to use it, without access to C++17?
c++, gcc, c++17, rhel
2
134
1
https://stackoverflow.com/questions/65690691/how-to-emulate-class-template-argument-deduction-pre-c17
59,281,720
How to read var/lib/rpm/Packages content
I have ' Packages ' file in path var/lib/rpm in my Centos:8 based image. How can I read/ get ' Packages ' file content As I read in some documentation that this file is of type berkeley DB.. not sure about this.
How to read var/lib/rpm/Packages content I have ' Packages ' file in path var/lib/rpm in my Centos:8 based image. How can I read/ get ' Packages ' file content As I read in some documentation that this file is of type berkeley DB.. not sure about this.
linux, centos, fedora, rhel, berkeley-db
2
4,444
3
https://stackoverflow.com/questions/59281720/how-to-read-var-lib-rpm-packages-content
50,107,982
RHE 7 Not respecting java Secure Random Seed
I have a quandary on my hands. I created an AES service to encrypt/decrypt sensitive information. The AES key is randomly generated using java's SecureRandom . I have a protected file that stores the seed and upon calling the service the seed is populated into the Secure Random class. To make sure it works I have the following logic: private boolean secureRandom(final String seed) { SecureRandom sr1 = new SecureRandom(seed.getBytes(UTF8_CHARSET)); SecureRandom sr2 = new SecureRandom(seed.getBytes(UTF8_CHARSET)); //Two secure random with the same seed should generate the same results boolean secureRandomWorks = sr1.nextLong() == sr2.nextLong(); if (!secureRandomWorks) { System.err.println("Secure random not supported. Defaulting to old key"); } return secureRandomWorks; } The idea here is I should be able to create two secure random objects with the same seed and they should both return the same value upon the call to nextLong() When I deploy my application on a windows machine this works fine, but when I deploy it on a RHEL 7 machine I get my error. I was under the impression that as long as the seed is the same, both instances will always produce the same output. This seems to be the case on windows, but not when I tested it on RHEL 7 this doesn't seem to be the case. I created this simple test to see verify: SecureRandom sr1 = new SecureRandom("encryptionKey".getBytes("UTF-8")); SecureRandom sr2 = new SecureRandom("encryptionKey".getBytes("UTF-8")); for (int i = 0; i < 1000; i++) { System.out.println(sr1.nextLong() == sr2.nextLong()); } And on windows every output was true while on RHEL 7 this was false. Any idea's suggestions on what might be causing RHEL 7 to ignore the seed?
RHE 7 Not respecting java Secure Random Seed I have a quandary on my hands. I created an AES service to encrypt/decrypt sensitive information. The AES key is randomly generated using java's SecureRandom . I have a protected file that stores the seed and upon calling the service the seed is populated into the Secure Random class. To make sure it works I have the following logic: private boolean secureRandom(final String seed) { SecureRandom sr1 = new SecureRandom(seed.getBytes(UTF8_CHARSET)); SecureRandom sr2 = new SecureRandom(seed.getBytes(UTF8_CHARSET)); //Two secure random with the same seed should generate the same results boolean secureRandomWorks = sr1.nextLong() == sr2.nextLong(); if (!secureRandomWorks) { System.err.println("Secure random not supported. Defaulting to old key"); } return secureRandomWorks; } The idea here is I should be able to create two secure random objects with the same seed and they should both return the same value upon the call to nextLong() When I deploy my application on a windows machine this works fine, but when I deploy it on a RHEL 7 machine I get my error. I was under the impression that as long as the seed is the same, both instances will always produce the same output. This seems to be the case on windows, but not when I tested it on RHEL 7 this doesn't seem to be the case. I created this simple test to see verify: SecureRandom sr1 = new SecureRandom("encryptionKey".getBytes("UTF-8")); SecureRandom sr2 = new SecureRandom("encryptionKey".getBytes("UTF-8")); for (int i = 0; i < 1000; i++) { System.out.println(sr1.nextLong() == sr2.nextLong()); } And on windows every output was true while on RHEL 7 this was false. Any idea's suggestions on what might be causing RHEL 7 to ignore the seed?
java, rhel, rhel7, secure-random
2
1,039
2
https://stackoverflow.com/questions/50107982/rhe-7-not-respecting-java-secure-random-seed
47,785,242
How to install optional rpms on RedHat 7?
I am trying to install RedHat optional rpms on RHEL 7 as follow. sudo yum install rhel-7-server-extras-rpms rhel-7-server-optional-rpms Loaded plugins: langpacks, product-id, rhnplugin, search-disabled-repos, subscription- : manager This system is receiving updates from RHN Classic or Red Hat Satellite. epel/x86_64/metalink | 13 kB 00:00:00 epel | 4.7 kB 00:00:00 nginx | 2.9 kB 00:00:00 nodesource | 2.5 kB 00:00:00 (1/5): epel/x86_64/group_gz | 266 kB 00:00:00 (2/5): epel/x86_64/updateinfo | 860 kB 00:00:00 (3/5): nginx/x86_64/primary_db | 31 kB 00:00:00 (4/5): nodesource/x86_64/primary_db | 29 kB 00:00:00 (5/5): epel/x86_64/primary_db | 6.1 MB 00:00:02 rhel-x86_64-server-7 | 1.5 kB 00:00:00 rhel-x86_64-server-7/group | 636 kB 00:00:00 rhel-x86_64-server-7/updateinfo | 2.1 MB 00:00:00 rhel-x86_64-server-7/primary | 25 MB 00:00:02 rhel-x86_64-server-7 17802/17802 No package rhel-7-server-extras-rpms available. No package rhel-7-server-optional-rpms available. Error: Nothing to do I am not able to understand this . Why are the optional rpms not available to me ? How do i setup a local repository of optional and extra rpms ?
How to install optional rpms on RedHat 7? I am trying to install RedHat optional rpms on RHEL 7 as follow. sudo yum install rhel-7-server-extras-rpms rhel-7-server-optional-rpms Loaded plugins: langpacks, product-id, rhnplugin, search-disabled-repos, subscription- : manager This system is receiving updates from RHN Classic or Red Hat Satellite. epel/x86_64/metalink | 13 kB 00:00:00 epel | 4.7 kB 00:00:00 nginx | 2.9 kB 00:00:00 nodesource | 2.5 kB 00:00:00 (1/5): epel/x86_64/group_gz | 266 kB 00:00:00 (2/5): epel/x86_64/updateinfo | 860 kB 00:00:00 (3/5): nginx/x86_64/primary_db | 31 kB 00:00:00 (4/5): nodesource/x86_64/primary_db | 29 kB 00:00:00 (5/5): epel/x86_64/primary_db | 6.1 MB 00:00:02 rhel-x86_64-server-7 | 1.5 kB 00:00:00 rhel-x86_64-server-7/group | 636 kB 00:00:00 rhel-x86_64-server-7/updateinfo | 2.1 MB 00:00:00 rhel-x86_64-server-7/primary | 25 MB 00:00:02 rhel-x86_64-server-7 17802/17802 No package rhel-7-server-extras-rpms available. No package rhel-7-server-optional-rpms available. Error: Nothing to do I am not able to understand this . Why are the optional rpms not available to me ? How do i setup a local repository of optional and extra rpms ?
rhel, rhel7, repo
2
24,656
2
https://stackoverflow.com/questions/47785242/how-to-install-optional-rpms-on-redhat-7
19,804,959
Where&#39;s python-argparse package in RHEL?
I'm using python's argparse package in CentOS6 . I installed it by the following command. # yum install python-argparse However, I could not find the package in RHEL5 and 6 . Where is it? How can I get argparse?
Where&#39;s python-argparse package in RHEL? I'm using python's argparse package in CentOS6 . I installed it by the following command. # yum install python-argparse However, I could not find the package in RHEL5 and 6 . Where is it? How can I get argparse?
python, rhel
2
12,314
2
https://stackoverflow.com/questions/19804959/wheres-python-argparse-package-in-rhel
18,676,559
Python - error 98 Adress already in use, how to make it faster? So that on kill and quick restart it does not fail?
I have a Python protocol running under CentOS 6.4/64-bit. Where i am having TCP server port 7007. In some cases like update new version or maintenance or on the fly restart to refresh buffer i need to restart the application as: server.py: class AServer(threading.Thread): def __init__(self, port): threading.Thread.__init__(self) self.port = port def run(self): host = '' s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) s.bind((host, self.port)) print bgcolors.BOOT s.listen(1) conn, addr = s.accept() print bgcolors.OK + 'contact', addr, 'on', self.now() while 1: try: data = conn.recv(1024) except socket.error: print bgcolors.OK + 'lost', addr, 'waiting..' s.listen(1) conn, addr = s.accept() print bgcolors.OK + 'contact', addr, 'on', self.now() continue if not data: ..... ... t = AServer(7007) t.start() Restart urgent on the fly (expecting to run within 1 second) but fails: $ ps aux | awk '/server.py/ {print $2}' | head -1 | xargs kill -9; $ nohup python /var/tmp/py-protocol/server.py & [root@IPSecVPN protocol]# python server.py Exception in thread Thread-1: Traceback (most recent call last): File "/usr/lib64/python2.6/threading.py", line 532, in __bootstrap_inner self.run() File "server.py", line 236, in run s.bind((host, self.port)) File "<string>", line 1, in bind error: [Errno 98] Address already in use
Python - error 98 Adress already in use, how to make it faster? So that on kill and quick restart it does not fail? I have a Python protocol running under CentOS 6.4/64-bit. Where i am having TCP server port 7007. In some cases like update new version or maintenance or on the fly restart to refresh buffer i need to restart the application as: server.py: class AServer(threading.Thread): def __init__(self, port): threading.Thread.__init__(self) self.port = port def run(self): host = '' s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) s.bind((host, self.port)) print bgcolors.BOOT s.listen(1) conn, addr = s.accept() print bgcolors.OK + 'contact', addr, 'on', self.now() while 1: try: data = conn.recv(1024) except socket.error: print bgcolors.OK + 'lost', addr, 'waiting..' s.listen(1) conn, addr = s.accept() print bgcolors.OK + 'contact', addr, 'on', self.now() continue if not data: ..... ... t = AServer(7007) t.start() Restart urgent on the fly (expecting to run within 1 second) but fails: $ ps aux | awk '/server.py/ {print $2}' | head -1 | xargs kill -9; $ nohup python /var/tmp/py-protocol/server.py & [root@IPSecVPN protocol]# python server.py Exception in thread Thread-1: Traceback (most recent call last): File "/usr/lib64/python2.6/threading.py", line 532, in __bootstrap_inner self.run() File "server.py", line 236, in run s.bind((host, self.port)) File "<string>", line 1, in bind error: [Errno 98] Address already in use
python, linux, centos, rhel
2
1,848
3
https://stackoverflow.com/questions/18676559/python-error-98-adress-already-in-use-how-to-make-it-faster-so-that-on-kill
16,247,845
dont understand cron. What is 56 11 * * * /usr/sbin/update-file.sh
What does this command mean in cron? How often will this run? When will it run? Will it run daily? 56 11 * * * /usr/sbin/update-file.sh
dont understand cron. What is 56 11 * * * /usr/sbin/update-file.sh What does this command mean in cron? How often will this run? When will it run? Will it run daily? 56 11 * * * /usr/sbin/update-file.sh
linux, unix, rhel
2
1,170
2
https://stackoverflow.com/questions/16247845/dont-understand-cron-what-is-56-11-usr-sbin-update-file-sh
71,934,306
Can&#39;t install powertools on RHEL 8.5
I'm new to Linux and I'm trying to setup my RHEL 8.5 VM machine. NAME="Red Hat Enterprise Linux VERSION="8.5 (Ootpa) I'm trying to install some scripts for course, and it is giving an error that PowerTools can't be found No matching repo to modify: powertools. I have tried several solution in other posts as follows: # dnf -y install dnf-plugins-core # dnf upgrade # dnf -y install [URL] # dnf config-manager --set-enabled powertools The first three commands execute and my system is updated but the last command is still having same error: Updating Subscription Management repositories. Error: No matching repo to modify: powertools. When I try to get repolist, it is not showing powertools # dnf repolist Updating Subscription Management repositories. repo id repo name BaseOS BaseOS appstream appstream codeready-builder-for-rhel-8-x86_64-rpms Red Hat CodeReady Linux Builder for RHEL 8 x86_64 (RPMs) epel Extra Packages for Enterprise Linux 8 - x86_64 epel-modular Extra Packages for Enterprise Linux Modular 8 - x86_64 rhel-8-for-x86_64-appstream-rpms Red Hat Enterprise Linux 8 for x86_64 - AppStream (RPMs) rhel-8-for-x86_64-baseos-rpms Red Hat Enterprise Linux 8 for x86_64 - BaseOS (RPMs) I hope someone can help. Thanks.
Can&#39;t install powertools on RHEL 8.5 I'm new to Linux and I'm trying to setup my RHEL 8.5 VM machine. NAME="Red Hat Enterprise Linux VERSION="8.5 (Ootpa) I'm trying to install some scripts for course, and it is giving an error that PowerTools can't be found No matching repo to modify: powertools. I have tried several solution in other posts as follows: # dnf -y install dnf-plugins-core # dnf upgrade # dnf -y install [URL] # dnf config-manager --set-enabled powertools The first three commands execute and my system is updated but the last command is still having same error: Updating Subscription Management repositories. Error: No matching repo to modify: powertools. When I try to get repolist, it is not showing powertools # dnf repolist Updating Subscription Management repositories. repo id repo name BaseOS BaseOS appstream appstream codeready-builder-for-rhel-8-x86_64-rpms Red Hat CodeReady Linux Builder for RHEL 8 x86_64 (RPMs) epel Extra Packages for Enterprise Linux 8 - x86_64 epel-modular Extra Packages for Enterprise Linux Modular 8 - x86_64 rhel-8-for-x86_64-appstream-rpms Red Hat Enterprise Linux 8 for x86_64 - AppStream (RPMs) rhel-8-for-x86_64-baseos-rpms Red Hat Enterprise Linux 8 for x86_64 - BaseOS (RPMs) I hope someone can help. Thanks.
centos, rhel, ef-power-tools
2
12,459
1
https://stackoverflow.com/questions/71934306/cant-install-powertools-on-rhel-8-5
37,525,884
ImportError: /lib/libc.so.6: version `GLIBC_2.14&#39; not found (required by /usr/lib/libpq.so.5)
I have to run a tornado webapp on a server using psycopg2 with postgresql as backend. I am working on Red Hat Enterprise Linux Server release 6.5. I set up a virtual environment. Now, on this server, i don't have root access, i can't use pip or yum and i can't copy text from the server(so i can't post the exact error log). I have been manually downloading tar files for the libraries and installing them. When i try to run the webapp, i get the following error: ImportError: /lib/libc.so.6: version `GLIBC_2.14' not found (required by /usr/lib/libpq.so.5) The version for libc.so.6 is 2.12. The error is displayed in psycopg2/ init .py. I apologize as this is not much to go on, but the server has a lot of restrictions. Any help will be appreciated.
ImportError: /lib/libc.so.6: version `GLIBC_2.14&#39; not found (required by /usr/lib/libpq.so.5) I have to run a tornado webapp on a server using psycopg2 with postgresql as backend. I am working on Red Hat Enterprise Linux Server release 6.5. I set up a virtual environment. Now, on this server, i don't have root access, i can't use pip or yum and i can't copy text from the server(so i can't post the exact error log). I have been manually downloading tar files for the libraries and installing them. When i try to run the webapp, i get the following error: ImportError: /lib/libc.so.6: version `GLIBC_2.14' not found (required by /usr/lib/libpq.so.5) The version for libc.so.6 is 2.12. The error is displayed in psycopg2/ init .py. I apologize as this is not much to go on, but the server has a lot of restrictions. Any help will be appreciated.
postgresql, psycopg2, glibc, rhel, libpq
2
24,016
1
https://stackoverflow.com/questions/37525884/importerror-lib-libc-so-6-version-glibc-2-14-not-found-required-by-usr-li
26,158,313
prevent rpmbuild from reading ~/.rpmmacros
Is there a way to prevent the rpmbuild command from reading the ~/.rpmmacros file? I want to add rpmbuild to a Make target and want to get a consistent build under any user account, regardless of the options somebody might have personally set in the rpmmacros file in his home directory. How can I do this? Or, is that actually a bad idea?
prevent rpmbuild from reading ~/.rpmmacros Is there a way to prevent the rpmbuild command from reading the ~/.rpmmacros file? I want to add rpmbuild to a Make target and want to get a consistent build under any user account, regardless of the options somebody might have personally set in the rpmmacros file in his home directory. How can I do this? Or, is that actually a bad idea?
linux, rpm, rhel, rpmbuild, building
2
428
2
https://stackoverflow.com/questions/26158313/prevent-rpmbuild-from-reading-rpmmacros
22,366,064
Sudo NOPASSWD not registering/working/showing
Sudo version: 1.7.2p1 I have the following (sanitised) lines in my sudoers file (at the end of the file): Runas_Alias GROUPALIASNAME = user1,user2, user3 Cmnd_Alias APP = /path/to/binary/setup -f *, /path/to/binary/patchbinary %groupname ALL=(root)NOPASSWD:SOFTWARE,(GROUPALIASNAME)NOPASSWD:APP If I run "sudo -ll" as a user from %groupname (user3) I get the following: Sudoers entry: RunAsUsers: root Commands: NOPASSWD: /bin/rpm, /usr/bin/up2date, /usr/bin/yum RunAsUsers: user1, user2, user3 Commands: /path/to/binary/setup -f *, /path/to/binary/patchbinary I can't seem to get the second "nopasswd" entry to work. I've tried removing all whitespace, adding whitespace. I've tried "NoPasswd:all" which just makes it give "all" rather than "nopasswd:all". I know the rule is matching because clearly I can see the very specific entries expanding. The entry being at the end should rule out any conflicting rules. Any ideas or suggestions on diagnosing this? Unfortunately its a 'secure' machine so applying patches/updating sudo isn't an option. **The reference to "SOFTWARE" is just a default sudoers entry. EDIT: For now I have worked around the issue by splitting the 1 line into 2 lines. I made no other modifications than to insert a line break with %groupname ALL= and it now correctly shows the right to execute the command with no password. However it still requires a password.
Sudo NOPASSWD not registering/working/showing Sudo version: 1.7.2p1 I have the following (sanitised) lines in my sudoers file (at the end of the file): Runas_Alias GROUPALIASNAME = user1,user2, user3 Cmnd_Alias APP = /path/to/binary/setup -f *, /path/to/binary/patchbinary %groupname ALL=(root)NOPASSWD:SOFTWARE,(GROUPALIASNAME)NOPASSWD:APP If I run "sudo -ll" as a user from %groupname (user3) I get the following: Sudoers entry: RunAsUsers: root Commands: NOPASSWD: /bin/rpm, /usr/bin/up2date, /usr/bin/yum RunAsUsers: user1, user2, user3 Commands: /path/to/binary/setup -f *, /path/to/binary/patchbinary I can't seem to get the second "nopasswd" entry to work. I've tried removing all whitespace, adding whitespace. I've tried "NoPasswd:all" which just makes it give "all" rather than "nopasswd:all". I know the rule is matching because clearly I can see the very specific entries expanding. The entry being at the end should rule out any conflicting rules. Any ideas or suggestions on diagnosing this? Unfortunately its a 'secure' machine so applying patches/updating sudo isn't an option. **The reference to "SOFTWARE" is just a default sudoers entry. EDIT: For now I have worked around the issue by splitting the 1 line into 2 lines. I made no other modifications than to insert a line break with %groupname ALL= and it now correctly shows the right to execute the command with no password. However it still requires a password.
bash, sudo, rhel
2
1,453
2
https://stackoverflow.com/questions/22366064/sudo-nopasswd-not-registering-working-showing
18,604,402
where is daemon command in rhel?
I write a service script following service script but I found there no daemon command, and I can't google how to install it
where is daemon command in rhel? I write a service script following service script but I found there no daemon command, and I can't google how to install it
linux, daemon, rhel
2
8,194
1
https://stackoverflow.com/questions/18604402/where-is-daemon-command-in-rhel
6,014,332
Issue with &#39;gem install sqlite3-ruby&#39; on RHEL6
I'm attempting to install Ruby on Rails 3 on RHEL 6. I've 'yum install' sqlite and sqlite-devel (which is sqlite3 on RHEL6). As you can see, sqlite3.h does exist where it needs to be. 'gem install sqlite3-ruby' still doesn't work, for some reason. SQLite3: [root@rhel6base gems]# yum list sqlite-devel Installed Packages sqlite-devel.x86_64 3.6.20-1.el6 @local Available Packages sqlite-devel.i686 3.6.20-1.el6 local [root@rhel6base gems]# ls -l /usr/include/sqlite3.h -rw-r--r--. 1 root root 268966 Nov 24 2009 /usr/include/sqlite3.h Gem install: [root@rhel6base gems]# gem install sqlite3-ruby Building native extensions. This could take a while... ERROR: Error installing sqlite3-ruby: ERROR: Failed to build gem native extension. /usr/bin/ruby extconf.rb checking for sqlite3.h... no sqlite3.h is missing. Try 'port install sqlite3 +universal' or 'yum install sqlite3-devel' and check your shared library search path (the location where your sqlite3 shared library is located). *** extconf.rb failed *** Could not create Makefile due to some reason, probably lack of necessary libraries and/or headers. Check the mkmf.log file for more details. You may need configuration options. Provided configuration options: --with-opt-dir --without-opt-dir --with-opt-include --without-opt-include=${opt-dir}/include --with-opt-lib --without-opt-lib=${opt-dir}/lib --with-make-prog --without-make-prog --srcdir=. --curdir --ruby=/usr/bin/ruby --with-sqlite3-dir --without-sqlite3-dir --with-sqlite3-include --without-sqlite3-include=${sqlite3-dir}/include --with-sqlite3-lib --without-sqlite3-lib=${sqlite3-dir}/lib Gem files will remain installed in /usr/lib64/ruby/gems/1.8/gems/sqlite3-1.3.3 for inspection. Results logged to /usr/lib64/ruby/gems/1.8/gems/sqlite3-1.3.3/ext/sqlite3/gem_make.out PS: It'd sure be nice to create a 'rhel6' tag, but I don't have enough reputation.
Issue with &#39;gem install sqlite3-ruby&#39; on RHEL6 I'm attempting to install Ruby on Rails 3 on RHEL 6. I've 'yum install' sqlite and sqlite-devel (which is sqlite3 on RHEL6). As you can see, sqlite3.h does exist where it needs to be. 'gem install sqlite3-ruby' still doesn't work, for some reason. SQLite3: [root@rhel6base gems]# yum list sqlite-devel Installed Packages sqlite-devel.x86_64 3.6.20-1.el6 @local Available Packages sqlite-devel.i686 3.6.20-1.el6 local [root@rhel6base gems]# ls -l /usr/include/sqlite3.h -rw-r--r--. 1 root root 268966 Nov 24 2009 /usr/include/sqlite3.h Gem install: [root@rhel6base gems]# gem install sqlite3-ruby Building native extensions. This could take a while... ERROR: Error installing sqlite3-ruby: ERROR: Failed to build gem native extension. /usr/bin/ruby extconf.rb checking for sqlite3.h... no sqlite3.h is missing. Try 'port install sqlite3 +universal' or 'yum install sqlite3-devel' and check your shared library search path (the location where your sqlite3 shared library is located). *** extconf.rb failed *** Could not create Makefile due to some reason, probably lack of necessary libraries and/or headers. Check the mkmf.log file for more details. You may need configuration options. Provided configuration options: --with-opt-dir --without-opt-dir --with-opt-include --without-opt-include=${opt-dir}/include --with-opt-lib --without-opt-lib=${opt-dir}/lib --with-make-prog --without-make-prog --srcdir=. --curdir --ruby=/usr/bin/ruby --with-sqlite3-dir --without-sqlite3-dir --with-sqlite3-include --without-sqlite3-include=${sqlite3-dir}/include --with-sqlite3-lib --without-sqlite3-lib=${sqlite3-dir}/lib Gem files will remain installed in /usr/lib64/ruby/gems/1.8/gems/sqlite3-1.3.3 for inspection. Results logged to /usr/lib64/ruby/gems/1.8/gems/sqlite3-1.3.3/ext/sqlite3/gem_make.out PS: It'd sure be nice to create a 'rhel6' tag, but I don't have enough reputation.
ruby, sqlite, rubygems, rhel, sqlite3-ruby
2
4,332
2
https://stackoverflow.com/questions/6014332/issue-with-gem-install-sqlite3-ruby-on-rhel6
62,189,834
pip download + why pip not download the latest version
we want to download the latest version of cffi pkg we do the following pip download cffi Collecting cffi Using cached [URL] Saved ./cffi-1.14.0-cp27-cp27mu-manylinux1_x86_64.whl Collecting pycparser (from cffi) Using cached [URL] Saved ./pycparser-2.20-py2.py3-none-any.whl Successfully downloaded cffi pycparser but we get - version - cffi-1.14.0 , and not the latest that is v1.8.3 ( [URL] ) ls -ltr total 492 -rw-r--r-- 1 root root 387834 Jun 4 08:04 cffi-1.14.0-cp27-cp27mu-manylinux1_x86_64.whl -rw-r--r-- 1 root root 112041 Jun 4 08:04 pycparser-2.20-py2.py3-none-any.whl so is it possible some way to download by pip the latest version of - cffi ? from - [URL] What’s New v1.8.3 When passing a void * argument to a function with a different pointer type, or vice-versa, the cast occurs automatically, like in C. The same occurs for initialization with ffi.new() and a few other places. However, I thought that char * had the same property—but I was mistaken. In C you get the usual warning if you try to give a char * to a char ** argument, for example. Sorry about the confusion. This has been fixed in CFFI by giving for now a warning, too. It will turn into an error in a future version.
pip download + why pip not download the latest version we want to download the latest version of cffi pkg we do the following pip download cffi Collecting cffi Using cached [URL] Saved ./cffi-1.14.0-cp27-cp27mu-manylinux1_x86_64.whl Collecting pycparser (from cffi) Using cached [URL] Saved ./pycparser-2.20-py2.py3-none-any.whl Successfully downloaded cffi pycparser but we get - version - cffi-1.14.0 , and not the latest that is v1.8.3 ( [URL] ) ls -ltr total 492 -rw-r--r-- 1 root root 387834 Jun 4 08:04 cffi-1.14.0-cp27-cp27mu-manylinux1_x86_64.whl -rw-r--r-- 1 root root 112041 Jun 4 08:04 pycparser-2.20-py2.py3-none-any.whl so is it possible some way to download by pip the latest version of - cffi ? from - [URL] What’s New v1.8.3 When passing a void * argument to a function with a different pointer type, or vice-versa, the cast occurs automatically, like in C. The same occurs for initialization with ffi.new() and a few other places. However, I thought that char * had the same property—but I was mistaken. In C you get the usual warning if you try to give a char * to a char ** argument, for example. Sorry about the confusion. This has been fixed in CFFI by giving for now a warning, too. It will turn into an error in a future version.
python, python-2.7, pip, rhel
2
90
2
https://stackoverflow.com/questions/62189834/pip-download-why-pip-not-download-the-latest-version
57,548,227
How to use -regex in the find command on Linux command line
The objective is to find and list anything with "messages" and/or "error.log" etc.. in the beginning then list both "messages.1..99" and "error.log.1..99" using regular expressions. This command works for however, it would require me to make many -or searches, but to simplify, I would like to have multiple in a set within the search. Like for instance: # find /var/log -maxdepth 1 -type f -size +1M -name [messages|error.log|secure.log|kern.log...]?[0-9]|[0-9][0-9] ! -iname "*.gz" not # find /var/log -maxdepth 1 -type f -size +1M -name "messages?[0-9]" -o -name "messages?[0-9][0-9]" How might I perform this command with regular expressions? # find /var/log -maxdepth 1 -type f -size +1M -name "[messages,error.log,kern,secure]?[0-9]" ! -iname "*.gz" My attempt with regex doesn't print anything in standard out: # find /var/log -maxdepth 1 -type f -size +1M -regex -name "[messages,error,kern,secure]?[0-9]" ! -iname "*.gz"
How to use -regex in the find command on Linux command line The objective is to find and list anything with "messages" and/or "error.log" etc.. in the beginning then list both "messages.1..99" and "error.log.1..99" using regular expressions. This command works for however, it would require me to make many -or searches, but to simplify, I would like to have multiple in a set within the search. Like for instance: # find /var/log -maxdepth 1 -type f -size +1M -name [messages|error.log|secure.log|kern.log...]?[0-9]|[0-9][0-9] ! -iname "*.gz" not # find /var/log -maxdepth 1 -type f -size +1M -name "messages?[0-9]" -o -name "messages?[0-9][0-9]" How might I perform this command with regular expressions? # find /var/log -maxdepth 1 -type f -size +1M -name "[messages,error.log,kern,secure]?[0-9]" ! -iname "*.gz" My attempt with regex doesn't print anything in standard out: # find /var/log -maxdepth 1 -type f -size +1M -regex -name "[messages,error,kern,secure]?[0-9]" ! -iname "*.gz"
linux, bash, shell, rhel
2
2,025
1
https://stackoverflow.com/questions/57548227/how-to-use-regex-in-the-find-command-on-linux-command-line
54,685,816
Scylla Installation issue on RHEL and CentOS
I am following RHEL instillation steps as decribed here [URL] Created and added bunch of RPM files in my local repo /etc/yum/repos.d/scylla.repo While installing below errors are coming. --> Processing Dependency: /usr/bin/python3 for package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 --> Finished Dependency Resolution Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: /usr/bin/python3 Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python-requests Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python34 Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libhwloc.so.5()(64bit) Error: Package: scylla-tools-core-2.1.6-20180701.fadba0b27.el7.noarch (scylla2.16) Requires: java-headless Error: Package: scylla-tools-2.1.6-20180701.fadba0b27.el7.noarch (scylla2.16) Requires: scylla-conf Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libcryptopp.so.6()(64bit) Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: PyYAML Error: Package: scylla-tools-core-2.1.6-20180701.fadba0b27.el7.noarch (scylla2.16) Requires: java-1.8.0-openjdk-headless Error: Package: scylla-jmx-2.1.6-20180701.315934b.el7.noarch (scylla2.16) Requires: java-1.8.0-openjdk-headless Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python3-pyudev Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libthrift-0.9.1.so()(64bit) Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python34-PyYAML Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libjsoncpp.so.0()(64bit) Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: hwloc Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python-setuptools Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libyaml-cpp.so.0.5()(64bit) Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python-urwid Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: collectd Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libgnutlsxx.so.28()(64bit) Error: Package: gcc-gnat-4.8.5-36.el7.x86_64 (@RHEL7) Requires: libgnat = 4.8.5-36.el7 Removing: libgnat-4.8.5-36.el7.x86_64 (@RHEL7) libgnat = 4.8.2-16.el7 libgnat = 4.8.5-36.el7 Obsoleted By: scylla-libgcc72-7.2.1-1.el7.centos.x86_64 (scylla2.16) Not found Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libprotobuf.so.8()(64bit) Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: scylla-conf Error: Package: scylla-tools-core-2.1.6-20180701.fadba0b27.el7.noarch (scylla2.16) Requires: python-yaml Please help to resolve this issue.
Scylla Installation issue on RHEL and CentOS I am following RHEL instillation steps as decribed here [URL] Created and added bunch of RPM files in my local repo /etc/yum/repos.d/scylla.repo While installing below errors are coming. --> Processing Dependency: /usr/bin/python3 for package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 --> Finished Dependency Resolution Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: /usr/bin/python3 Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python-requests Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python34 Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libhwloc.so.5()(64bit) Error: Package: scylla-tools-core-2.1.6-20180701.fadba0b27.el7.noarch (scylla2.16) Requires: java-headless Error: Package: scylla-tools-2.1.6-20180701.fadba0b27.el7.noarch (scylla2.16) Requires: scylla-conf Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libcryptopp.so.6()(64bit) Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: PyYAML Error: Package: scylla-tools-core-2.1.6-20180701.fadba0b27.el7.noarch (scylla2.16) Requires: java-1.8.0-openjdk-headless Error: Package: scylla-jmx-2.1.6-20180701.315934b.el7.noarch (scylla2.16) Requires: java-1.8.0-openjdk-headless Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python3-pyudev Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libthrift-0.9.1.so()(64bit) Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python34-PyYAML Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libjsoncpp.so.0()(64bit) Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: hwloc Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python-setuptools Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libyaml-cpp.so.0.5()(64bit) Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: python-urwid Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: collectd Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libgnutlsxx.so.28()(64bit) Error: Package: gcc-gnat-4.8.5-36.el7.x86_64 (@RHEL7) Requires: libgnat = 4.8.5-36.el7 Removing: libgnat-4.8.5-36.el7.x86_64 (@RHEL7) libgnat = 4.8.2-16.el7 libgnat = 4.8.5-36.el7 Obsoleted By: scylla-libgcc72-7.2.1-1.el7.centos.x86_64 (scylla2.16) Not found Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: libprotobuf.so.8()(64bit) Error: Package: scylla-server-2.1.6-0.20180701.7d2150a05.el7.x86_64 (scylla2.16) Requires: scylla-conf Error: Package: scylla-tools-core-2.1.6-20180701.fadba0b27.el7.noarch (scylla2.16) Requires: python-yaml Please help to resolve this issue.
cassandra, rhel, rhel7, scylla
2
845
1
https://stackoverflow.com/questions/54685816/scylla-installation-issue-on-rhel-and-centos
38,735,225
typeerror: __init__() got an unexpected keyword argument &#39;timeout&#39; pymongo
I'm trying to extract the facebook data into mongoDB. I'm using python 2.7.3 and pymongo-3.3.0 on the linux environment (RHEL), while extracting the data, I got the following error. Exception AttributeError: "'Cursor' object has no attribute '_Cursor__id'" in <bound method Cursor.__del__ of <pymongo.cursor.Cursor object at 0x48fa110>> ignored (<type 'exceptions.TypeError'>, TypeError("__init__() got an unexpected keyword argument 'timeout'",),<traceback object at 0x490a638>) Please suggest me how to fix this.
typeerror: __init__() got an unexpected keyword argument &#39;timeout&#39; pymongo I'm trying to extract the facebook data into mongoDB. I'm using python 2.7.3 and pymongo-3.3.0 on the linux environment (RHEL), while extracting the data, I got the following error. Exception AttributeError: "'Cursor' object has no attribute '_Cursor__id'" in <bound method Cursor.__del__ of <pymongo.cursor.Cursor object at 0x48fa110>> ignored (<type 'exceptions.TypeError'>, TypeError("__init__() got an unexpected keyword argument 'timeout'",),<traceback object at 0x490a638>) Please suggest me how to fix this.
linux, python-2.7, pymongo, rhel
2
2,014
1
https://stackoverflow.com/questions/38735225/typeerror-init-got-an-unexpected-keyword-argument-timeout-pymongo
38,604,162
erlang is no longer supported by Red Hat Enterprise Linux?
I'm trying installing Erlang on my REHL7.2, but unfortunately I failed. I tried two approaches First I tried downloading a RPM package from official site of erlang (in fact the rpm package is for CentOS, but there's no RPM package for RHEL anyway), but error: Failed dependencies: libwx_baseu-2.8.so.0()(64bit) is needed by esl-erlang-19.0-1.x86_64 libwx_baseu-2.8.so.0(WXU_2.8)(64bit) is needed by esl-erlang-19.0-1.x86_64 Then I tried $ wget [URL] $ rpm -Uvh erlang-solutions-1.0-1.noarch.rpm But it failed too because No package erlang available. Error: Nothing to do I wonder whether Erlang is no longer supported by RHEL? Otherwise how can I install Erlang on my RHEL7.2?
erlang is no longer supported by Red Hat Enterprise Linux? I'm trying installing Erlang on my REHL7.2, but unfortunately I failed. I tried two approaches First I tried downloading a RPM package from official site of erlang (in fact the rpm package is for CentOS, but there's no RPM package for RHEL anyway), but error: Failed dependencies: libwx_baseu-2.8.so.0()(64bit) is needed by esl-erlang-19.0-1.x86_64 libwx_baseu-2.8.so.0(WXU_2.8)(64bit) is needed by esl-erlang-19.0-1.x86_64 Then I tried $ wget [URL] $ rpm -Uvh erlang-solutions-1.0-1.noarch.rpm But it failed too because No package erlang available. Error: Nothing to do I wonder whether Erlang is no longer supported by RHEL? Otherwise how can I install Erlang on my RHEL7.2?
erlang, rhel
2
7,774
4
https://stackoverflow.com/questions/38604162/erlang-is-no-longer-supported-by-red-hat-enterprise-linux
34,510,182
Error: package does not exist. Is the issue code related?
I'm trying to compile Java code in RHEL 7 and i get this kind of errors error: package org.apache.xml.serialize does not exist import org.apache.xml.serialize.Method; symbol: class SistPrint location: class CrearSocketMIE ./GRP/Comunicaciones/ComMIE/CrearSocketMIE.java:187: error: cannot find symbol public SocketCliente getSocketMIE(int lan) { and so on, when i try in Ubuntu i don't get them... what am i seeing? is this code related or is the SO environment?. Both have the same jdk version (jdk1.8.0_65), how can they behave differently? What config am i missing (sure JAVA_HOME is already set) is there anything else?, I have a Make file that does different things depending the environment, really strange. What do you think?
Error: package does not exist. Is the issue code related? I'm trying to compile Java code in RHEL 7 and i get this kind of errors error: package org.apache.xml.serialize does not exist import org.apache.xml.serialize.Method; symbol: class SistPrint location: class CrearSocketMIE ./GRP/Comunicaciones/ComMIE/CrearSocketMIE.java:187: error: cannot find symbol public SocketCliente getSocketMIE(int lan) { and so on, when i try in Ubuntu i don't get them... what am i seeing? is this code related or is the SO environment?. Both have the same jdk version (jdk1.8.0_65), how can they behave differently? What config am i missing (sure JAVA_HOME is already set) is there anything else?, I have a Make file that does different things depending the environment, really strange. What do you think?
java, linux, ubuntu, rhel
2
2,213
1
https://stackoverflow.com/questions/34510182/error-package-does-not-exist-is-the-issue-code-related