Claudiu Demian wrote on 26/11/10 18:36:
> I found a problem with installing the package that I compiled on some of
> the nodes:
> wn59 ~]# wget
> http://ui01.mosigrid.utcluj.ro/~demi/openmpi-1.4.3-1.x86_64.rpm&& rpm
> -Uvh openmpi-1.4.3-1.x86_64.rpm&& rm -rf openmpi-1.4.3-1.x86_64.rpm
> --2010-11-26 17:29:29--
> http://ui01.mosigrid.utcluj.ro/~demi/openmpi-1.4.3-1.x86_64.rpm
> Resolving ui01.mosigrid.utcluj.ro... 217.73.173.20
> Connecting to ui01.mosigrid.utcluj.ro|217.73.173.20|:80... connected.
> HTTP request sent, awaiting response... 200 OK
> Length: 8649982 (8.2M) [application/x-rpm]
> Saving to: `openmpi-1.4.3-1.x86_64.rpm.1'
>
> 100%[===================================================================================================================>]
> 8,649,982 --.-K/s in 0.09s
>
> 2010-11-26 17:29:29 (92.5 MB/s) - `openmpi-1.4.3-1.x86_64.rpm.1' saved
> [8649982/8649982]
>
> error: Failed dependencies:
> openmpi = 1.4-4.el5 is needed by (installed) openmpi-devel-1.4-4.el5.x86_64
>
> Is openmpi-devel needed anymore? Or should I try and create a package
> for that too? Does mpicc come with the normal package or the -devel one?
It looks like you have openmpi = 1.4-4.el5 and openmpi-devel-1.4-4.el5
already installed on your WNs. I think first you need to remove them and
then install the custom built one.
Nikolay.
>
> Thanks,
> Claudiu
>
> On 11/26/2010 05:04 PM, Nikolay Kutovskiy wrote:
>> Claudiu Demian wrote on 26/11/10 17:59:
>>> Hi Enol,
>>>
>>> On 11/26/2010 04:52 PM, Enol Fernández del Castillo wrote:
>>>> Hi Claudiu,
>>>>
>>>> On 11/26/2010 03:12 PM, Claudiu Demian wrote:
>>>>> Hi Enol,
>>>>>
>>>>> You were right, I had that line.
>>>>> I've changed it according to your indications. This is my output now:
>>>>>
>>>>> # echo "#PBS -l nodes=64" | perl /var/spool/pbs/submit_filter.pl
>>>>> #PBS -l nodes=1:ppn=16+48:ppn=1
>>>>
>>>> Still does not look right, that's one node with 16 cpus and 48 with 1,
>>>> but in a previous mail you were saying that you have only 4 nodes with
>>>> 16 cpus each. So the expression should be like nodes=4:ppn=16.
>>>
>>> At our cluster, we have 64 nodes with 16 CPUs each. I only wanted to use
>>> 4 of the nodes with all of their 64 CPUs. I am sorry for the confusion.
>>> I guess the output should be nodes=64:ppn=16. (nodes=63 actually, since
>>> one of them is in service).
>>>
>>>>
>>>>>
>>>>> If I try to run a job now, it exists with a compiling error:
>>>>>
>>>>> -<START PRE-RUN
>>>>> HOOK>---------------------------------------------------
>>>>> Compiling /home/ops053/home_cream_266664921/CREAM266664921/hello
>>>>> mpicc -o /home/ops053/home_cream_266664921/CREAM266664921/hello
>>>>> /home/ops053/home_cream_266664921/CREAM266664921/hello.c
>>>>> Error compiling program. Exiting...
>> http://ui01.mosigrid.utcluj.ro/~demi/demi__eje7VNQJnWbhtJpZdPPvQ/mpi-start.err
>> :
>> + echo '-<START PRE-RUN
>> HOOK>---------------------------------------------------'
>> + pre_run_hook
>> + echo 'Compiling /home/ops053/home_cream_266664921/CREAM266664921/hello'
>> + cmd='mpicc -o /home/ops053/home_cream_266664921/CREAM266664921/hello
>> /home/ops053/home_cream_266664921/CREAM266664921/hello.c'
>> + echo mpicc -o /home/ops053/home_cream_266664921/CREAM266664921/hello
>> /home/ops053/home_cream_266664921/CREAM266664921/hello.c
>> + mpicc -o /home/ops053/home_cream_266664921/CREAM266664921/hello
>> /home/ops053/home_cream_266664921/CREAM266664921/hello.c
>> mpi-hooks.sh: line 15: mpicc: command not found
>> + '[' '!' 127 -eq 0 ']'
>> + echo 'Error compiling program. Exiting...'
>>
>> Make sure that mpicc is in the $PATH.
>>
>> Nikolay.
>>
|