[Ekhi-users] Ekhi SLURM configuration

Inigo Aldazabal Mensa inigo.aldazabalm at ehu.eus
Wed Nov 4 21:22:38 CET 2020


Hi all,

I finally got to try Slurm job accounting changes in a test environment
and all jobs and user allocations seemed unaffected.

That being said, I'll like to be sure that you are not running any
important, deadline kind of job, just in case jobs get canceled, even
if they shouldn't.

So, if no one objects, my plan is to implement the changes in Slurm next
Monday the 9th. Speak now or forever hold your peace :-)

Bests,

Iñigo

-- 
Iñigo Aldazabal Mensa, Ph.D.
Computing Service Manager / Scientific Computing Specialist
Centro de Física de Materiales (CSIC, UPV/EHU)
Paseo Manuel de Lardizabal, 5
20018 San Sebastian - Guipuzcoa
SPAIN

phone: +34-943-01-8780
e-mail: inigo.aldazabalm en ehu.eus inigo.aldazabal en csic.es
pgp key id: 0xDBCC8369





On Thu, 29 Oct 2020 09:56:17 +0100
Inigo Aldazabal Mensa <inigo.aldazabalm en ehu.eus> wrote:

> On Thu, 29 Oct 2020 09:35:48 +0100
> Ion Errea <ion.errea en ehu.eus> wrote:
> 
> > Hi Iñigo,
> > 
> > I agree with that change. So you can go on  
> 
> Done, let's see.
> 
> I expect to check today the major changes in a test environment I'm
> setting up for this. I'll report back about the implications on the
> present setup.
> 
> Iñigo
> 
> .
> > 
> > Let’s see if that improves the performance.
> > 
> > Bests,
> > 
> > Ion Errea
> > 
> > Fisika Aplikatua 1 saila, Gipuzkoako Ingeniaritza Eskola, and
> > Centro de Física de Materiales (CSIC-UPV/EHU),
> > University of the Basque Country (UPV/EHU)
> >          
> > Manuel de Lardizabal 5, 20018 Donostia, 
> > Basque Country, Spain
> > 
> > Tel:      +34 943 01 8417
> > Email:  ion.errea en ehu.eus
> > Web:    http://ionerrea.wordpress.com/
> >   
> > > On 29 Oct 2020, at 07:37, Inigo Aldazabal Mensa
> > > <inigo.aldazabalm en ehu.eus> wrote:
> > > 
> > > Hi Antonella,
> > > 
> > > I'm testing the queue system changes as it has implications in the
> > > present setup that I have to check out first. It is a major change
> > > in the basics of the accounting and scheduling system, and I want
> > > to coordinate with all of you in case the running jobs and present
> > > accounts are affected because of it.
> > > 
> > > That being sayd, removing the higher priority for small jobs is an
> > > independent, minor change  and, should Ion agree, can be done
> > > straight on.
> > > 
> > > Bests,
> > > 
> > > Iñigo
> > > 
> > > 
> > > 
> > > 
> > > On Wed, 28 Oct 2020 16:59:30 +0100
> > > Antonella Meninno <ameninno001 en ikasle.ehu.es> wrote:
> > >     
> > >> Hi Iñigo,
> > >> 
> > >> I'd like to know if there has been any change to the queue
> > >> system, and how it works now. In particular, I'd like to know if
> > >> the absolute priority to small jobs has been removed.
> > >> 
> > >> I would say that giving absolute priority to small jobs presents
> > >> an issue, the large jobs get stuck and never finish. I think
> > >> that this should change, in order to give space to large jobs
> > >> too.
> > >> 
> > >> Bests,
> > >> 
> > >> Antonella Meninno
> > >> 
> > >> On Mon, Oct 19, 2020 at 1:50 PM Inigo Aldazabal Mensa
> > >> <inigo.aldazabalm en ehu.eus> wrote:
> > >>     
> > >>> Hi Martin,
> > >>> 
> > >>> (General note: please start a new thread for new topics, do not
> > >>> follow old ones!)
> > >>> 
> > >>> Yes, I have to setup the queue system so that history is taken
> > >>> into account. I'll put this up in my todo list and try to do it
> > >>> along this week.
> > >>> 
> > >>> Technically I have to setup the "Slurm Accounting Database"
> > >>> which will allow us to use the all the features of the
> > >>> "Multifactor Prority Plugin", and also "Quality of Service"
> > >>> limits (see below)
> > >>> 
> > >>> https://slurm.schedmd.com/priority_multifactor.html
> > >>> 
> > >>> More below...
> > >>> 
> > >>> On Mon, 19 Oct 2020 10:47:01
> > >>> +0200 Martin Gutierrez <martin00gutierrez en gmail.com> wrote:
> > >>>     
> > >>>> Hi Iñigo, we identified some issues we the "free share" of the
> > >>>> queue after discussing with the group. We think they are
> > >>>> important and could be improved.
> > >>>> 
> > >>>> We would like to have the cluster usage history be taken into
> > >>>> account. Right now priority seems to be purely based on job
> > >>>> size (in terms of core-RAM usage). This means that a big
> > >>>> projects that can be divided into small pieces will be
> > >>>> blocking the cluster for all the others that need a single big
> > >>>> bulky job even if they are never using the cluster.      
> > >>> 
> > >>> Yes, now small jobs are prioritized (option
> > >>> PriorityFavorSmall=Yes)
> > >>> 
> > >>> https://slurm.schedmd.com/priority_multifactor.html#jobsize
> > >>>     
> > >>>> We would like that, for a single user, the jobs will be
> > >>>> executed in the order they have been listed in the queue when
> > >>>> possible. Since the queue is prioritizing small jobs, an user
> > >>>> that has to run long and short job will see the long jobs be
> > >>>> blocked by the prioritized small jobs. So, when possible for a
> > >>>> single user they should be ordered by the date they have been
> > >>>> listed into queue.      
> > >>> 
> > >>> Strict ordering I see very difficult to achieve, but *not*
> > >>> prioritizing small jobs is easy and just a matter of changing
> > >>> the previous option.     
> > >>>> Reserving resources when a job has been in top priority for a
> > >>>> long amount of time. Most of our jobs are less then one node. A
> > >>>> two nodes the job would get stuck until the two nodes are
> > >>>> freed. In our case the cluster is full most of the times
> > >>>> meaning the job would remain stuck in queue for a long time
> > >>>> even if prioritized.      
> > >>> 
> > >>> Slurm should be clever enough to leave an "empty" node waiting
> > >>> in order to fit a two node job. This should work mostly ok.
> > >>> Again not giving higher priority to small jobs can improve this.
> > >>>     
> > >>>> Each user should be allowed to four nodes for their long jobs
> > >>>> regardless of the amount of long jobs running already.
> > >>>> Althought we expected the cluster to receive mainly short
> > >>>> jobs, right now some of us are dealing with bulky and
> > >>>> expensive calculation all at the same time. If someone is then
> > >>>> using the four nodes, the rest of the users are forced to wait
> > >>>> even if the cluster is free.      
> > >>> 
> > >>> This kind of user limits can be applied (I have still to check
> > >>> exactly the level of detail that can be specified) using what in
> > >>> Slurm is called Quality of Service (QOS), but for this first I
> > >>> have to set up the Slurm Accounting Database.
> > >>> 
> > >>> https://slurm.schedmd.com/qos.html#limits
> > >>> 
> > >>> 
> > >>> Should you be interested you can take a look at ekhi slurm
> > >>> configuration file /etc/slurm/slurm.conf
> > >>> 
> > >>> In any case I'll talk to Ion about this ideas and come back
> > >>> with a better defined plan.
> > >>> 
> > >>> Bests,
> > >>> 
> > >>> Iñigo
> > >>> 
> > >>>     
> > >>>> On Aug 21 2020, at 10:32 pm, Inigo Aldazabal Mensa
> > >>>> <inigo.aldazabalm en ehu.eus> wrote:      
> > >>>>> Hi all,
> > >>>>> 
> > >>>>> It seems you are having some problems running jobs on ekhi.
> > >>>>> But some don't, as you can see with "squeue".
> > >>>>> 
> > >>>>> In order to troubleshoot your problems please report:
> > >>>>> 1.- Directory you did "sbatch" from
> > >>>>> 2.- slurm.script used
> > >>>>> 3.- JobID
> > >>>>> 
> > >>>>> ekhi11 had some problems and it should be fixed now.
> > >>>>> 
> > >>>>> Iñigo
> > >>>>> 
> > >>>>> On Wed, 19 Aug 2020
> > >>>>> 18:50:42 +0200 Inigo Aldazabal Mensa
> > >>>>> <inigo.aldazabalm en ehu.eus> wrote:      
> > >>>>>> Hi,
> > >>>>>> 
> > >>>>>> Again we had problems with the CFM Data Center 2 cooling
> > >>>>>> system that forced me to cancel all running jobs and shut
> > >>>>>> down all Ekhi computing nodes :-((
> > >>>>>> 
> > >>>>>> Currently I'm working remotely most of the time (in fact,
> > >>>>>> again, I was on a free day today) but tomorrow I'll go to the
> > >>>>>> CFM and talk to the cooling system technicians to check this.
> > >>>>>> Also, a new cooling system to avoid this problems is on the
> > >>>>>> way, but just a couple of weeks of being installed.
> > >>>>>> 
> > >>>>>> Depending on my talk with the technicians tomorrow I'll power
> > >>>>>> back on Ekhi computing nodes, but I can not say for sure.
> > >>>>>> I'll keep you informed.
> > >>>>>> 
> > >>>>>> Bests (to say something),
> > >>>>>> 
> > >>>>>> Iñigo
> > >>>>>>     
> > >>>>> _______________________________________________
> > >>>>> Ekhi-users mailing list
> > >>>>> Ekhi-users en list.ehu.eus
> > >>>>> http://list.ehu.eus/mailman/listinfo/ekhi-users
> > >>>>>     
> > >>>>     
> > >>> _______________________________________________
> > >>> Ekhi-users mailing list
> > >>> Ekhi-users en list.ehu.eus
> > >>> http://list.ehu.eus/mailman/listinfo/ekhi-users
> > >>>     
> > > _______________________________________________
> > > Ekhi-users mailing list
> > > Ekhi-users en list.ehu.eus
> > > http://list.ehu.eus/mailman/listinfo/ekhi-users    
> >   
> _______________________________________________
> Ekhi-users mailing list
> Ekhi-users en list.ehu.eus
> http://list.ehu.eus/mailman/listinfo/ekhi-users


More information about the Ekhi-users mailing list