gridengine | lightweight Python library for distributed computing | Job Scheduling library

 by   hbristow Python Version: Current License: BSD-3-Clause

kandi X-RAY | gridengine Summary

kandi X-RAY | gridengine Summary

gridengine is a Python library typically used in Data Processing, Job Scheduling applications. gridengine has no bugs, it has build file available, it has a Permissive License and it has low support. However gridengine has 2 vulnerabilities. You can download it from GitHub.

GridEngine streamlines the process of managing distributed computing on a Sun Grid Engine. It was designed for iterating over algorithm and experiment design on a computing cluster. GridEngine was intentionally designed to match the API of the built-in multiprocessing.Process and threading.Thread classes. If you have ever used these, the gridengine.Job class will be familiar to you. At its core, gridengine is designed to transparently schedule and execute `Job`s on a Sun Grid Engine computing cluster and return the results once the jobs have completed. All scheduling and communication whilst jobs are running are handled by gridengine.
Support
    Quality
      Security
        License
          Reuse

            kandi-support Support

              gridengine has a low active ecosystem.
              It has 21 star(s) with 7 fork(s). There are 2 watchers for this library.
              OutlinedDot
              It had no major release in the last 6 months.
              gridengine has no issues reported. There are 2 open pull requests and 0 closed requests.
              It has a neutral sentiment in the developer community.
              The latest version of gridengine is current.

            kandi-Quality Quality

              gridengine has no bugs reported.

            kandi-Security Security

              gridengine has 2 vulnerability issues reported (0 critical, 1 high, 1 medium, 0 low).

            kandi-License License

              gridengine is licensed under the BSD-3-Clause License. This license is Permissive.
              Permissive licenses have the least restrictions, and you can use them in most projects.

            kandi-Reuse Reuse

              gridengine releases are not available. You will need to build from source code and install.
              Build file is available. You can build the component from source.
              Installation instructions, examples and code snippets are available.

            Top functions reviewed by kandi - BETA

            kandi has reviewed gridengine and discovered the below as its top functions. This is intended to give you an instant insight into gridengine implemented functionality, and help decide if they suit your requirements.
            • Invoke the given function f
            • Dispatch the given list of jobs
            • Start the job
            • Run the target
            • Fetch data from grid
            • Store the result
            • Wait for all jobs to finish
            • Terminate all jobs
            • Create a partial function
            • Test if x is a number
            • Run a job from the command line
            • Read content of file
            Get all kandi verified functions for this library.

            gridengine Key Features

            No Key Features are available at this moment for gridengine.

            gridengine Examples and Code Snippets

            No Code Snippets are available at this moment for gridengine.

            Community Discussions

            QUESTION

            Extending Kaldi Aspire: bad variable error while Recompiling HCLG.fst using new lexicon and grammar files
            Asked 2020-Mar-27 at 14:23

            I have successfully setup and run the Kaldi Aspire recipe on my WSL. Now i was working on a POC where i want to extend the ASPIRE recipe by making a new corpus, dictionary, language model and merge it with the original HCLG.fst. I followed this blog post. I have been able to sucessfully create the new dictionary, language model and merged the input files. However i am getting the following error when i try to recompile the HCLG.fst with new Lexicon and grammar.

            ...

            ANSWER

            Answered 2020-Mar-27 at 14:23

            QUESTION

            Typo3 Fluid ignoring max image sizes set in typoscript
            Asked 2020-Feb-18 at 09:39

            I've set in my Constants

            ...

            ANSWER

            Answered 2020-Feb-17 at 16:00

            Short anwer:

            The image viewhelper does not automatically apply any dimensions defined in the constants. f:image is part of ext:fluid, the constants you set are part of ext:fluid_styled_content. The image viewhelper always needs a width or a maxWidth argument if you want to set one. You need to pass your constants to the frontend.

            Long answer:

            You can see, how it is done in fluid_styled_content:

            Have a look into: typo3/sysext/fluid_styled_content/Configuration/TypoScript/ContentElement/Image.typoscript A DataProcessor passes this constants to the GalleryProcessor

            Source https://stackoverflow.com/questions/60260840

            QUESTION

            Gridengine: error: commlib error: got select error (connection refused)
            Asked 2018-Dec-11 at 13:59

            I just installed gridengine & getting error when doing qstat:

            ...

            ANSWER

            Answered 2018-Dec-11 at 13:59

            Most of the setup issues went away when installing the Debian version. need version 8 since debian is ver 8 & only debian master works ( gridengine-master_8.1.9+dfsg-8_amd64.deb).

            Source https://stackoverflow.com/questions/52028613

            QUESTION

            superclass mismatch for class Queue (TypeError)
            Asked 2018-Mar-14 at 08:29

            When running the analyze.rb script from Univa Grid Engine Open Core , I get a TypeError:

            ...

            ANSWER

            Answered 2018-Mar-14 at 08:29

            The script adds a class Queue like this:

            Source https://stackoverflow.com/questions/49260040

            QUESTION

            'scl enable' not setting PATH with bash
            Asked 2018-Feb-05 at 13:08

            I'm trying to use software collections on CentOS 6.8 server, but it won't set the environment variable PATH corectly if the command passed is "bash", but "tcsh" works... (however we don't use tcsh on this machine)

            Example:

            ...

            ANSWER

            Answered 2017-May-12 at 23:23

            Thank you Dominic you were on to something. I originally checked ~/.bash* files as well as /etc/bash* and /etc/profile but after your comment, I found several scripts in /etc/profile.d/ that we being executed, and one of them set the PATH explicitly without appending. I added $PATH back in there and now scl enable is working as expected!

            Source https://stackoverflow.com/questions/43880358

            QUESTION

            Using pandas to analyzing a over 20G data frame, out of memory, when specifying chunksize still wouldn't work
            Asked 2017-Aug-05 at 23:59

            I have the following code to analyze a huge dataframe file (22G, over 2million rows and 3K columns). I tested the code in a smaller dataframe and it ran OK (head -1000 hugefile.txt). However, when I ran the code on the huge dataframe, it gave me "segmentation fault" core dump. It output a core.number binary file.

            I did some internet search and came up with using low_memory =False, and trying to read the DataFrame by defining chunksize=1000, iterator= True, and then pandas.concat the chunks, but this still gave me memory problem (core dump). It wouldn't even read in the entire file before the core dump as I tested just read the file and print some text. Please help and let me know if there are solutions that I can analyze this huge file.

            Version

            python version: 3.6.2
            numpy version: 1.13.1
            pandas version: 0.20.3
            OS: Linux/Unix

            Script

            ...

            ANSWER

            Answered 2017-Aug-05 at 23:09

            You aren't processing your data in chunks at all.

            With data1 = pd.read_csv('...', chunksize=10000, iterator=True), data1 becomes a pandas.io.parser.TextFileReader, which is an iterator that yields chunks of 10000 rows of your CSV data as a DataFrame.

            But then pd.concat consumes this entire iterator, and so attempts to load the whole CSV into memory, defeating the purpose of using chunksize and iterator entirely.

            Properly using chunksize and iterator

            In order to process your data in chunks, you have to iterate over the actual DataFrame chunks yielded by the iterator read.csv provides.

            Source https://stackoverflow.com/questions/45511085

            QUESTION

            Enabling Univa Grid Engine Resource Reservation without a time limit on jobs
            Asked 2017-Mar-15 at 19:43

            My organization has a server cluster running Univa Grid Engine 8.4.1, with users submitting various kinds of jobs, some using a single CPU core, and some using OpenMPI to utilize multiple cores, all with varying and unpredictable run-times.

            We've enabled a ticketing system so that one user can't hog the entire queue, but if the grid and queue are full of single-CPU jobs, no multi-CPU job can ever start (they just sit at the top of the queue waiting for the required number of cpu slots to become free, which generally never happens). We're looking to configure Resource Reservation such that, if the MPI job is the next in the queue, the grid will hold slots open as they become free until there's enough to submit the MPI job, rather than filling them with the single-CPU jobs that are further down in the queue.

            I've read (here for example) that the grid makes the decision of which slots to "reserve" based on how much time is remaining on the jobs running in those slots. The problem we have is that our jobs have unknown run-times. Some take a few seconds, some take weeks, and while we have a rough idea how long a job will take, we can never be sure. Thus, we don't want to start running qsub with hard and soft time limits through -l h_rt and -l s_rt, or else our jobs could be killed prematurely. Resource Reservation appears to be using the default_duration, which we set to infinity for lack of a better number to use, and treating all jobs equally. Its picking slots filled by month-long jobs which have already been running for a few days, instead of slots filled by minute-long jobs which have only been running for a few seconds.

            Is there a way to tell the scheduler to reserve slots for a multi-CPU MPI job as they become available, rather than pre-select slots based on some perceived run-time of the jobs in them?

            ...

            ANSWER

            Answered 2017-Mar-15 at 19:43

            Unfortunately I'm not aware of a way to do what you ask - I think that the reservation is created once at the time that the job is submitted, not progressively as slots become free. If you haven't already seen the design document for the Resource Reservation feature, it's worth a look to get oriented to the feature.

            Instead, I'm going to suggest some strategies for confidently setting job runtimes. The main problem when none of your jobs have runtimes is that Grid Engine can't reserve space infinitely in the future, so even if you set some really rough runtimes (within an order of magnitude of the true runtime), you may get some positive results.

            1. If you've run a similar job previously, one simple rule of thumb is to set max runtime to 150% of the typical or maximum runtime of the job, based on historical trends. Use qacct or parse the accounting file to get hard data. Of course, tweak that percentage to whatever suits your risk threshold.
            2. Another rule of thumb is to set the max runtime not based on the job's true runtime, but based on a sense around "after this date, the results won't be useful" or "if it takes this long, something's definitely wrong". If you need an answer by Friday, there's no sense in setting the runtime limit for three months out. Similarly, if you're running md5sum on typically megabyte-sized files, there's no sense in setting a 1-day runtime limit; those jobs ought to only take a few seconds or minutes, and if it's really taking a long time, then something is broken.
            3. If you really must allow true indefinite-length jobs, then one option is to divide your cluster into infinite and finite queues. Jobs specifying a finite runtime will be able to use both queues, while infinite jobs will have fewer resources available; this will incentivize users to work a little harder at picking runtimes, without forcing them to do so.

            Finally, be sure that the multi-slot jobs are submitted with the -R y qsub flag to enable the resource reservation system. This could go in the system default sge_request file, but that's generally not recommended as it can reduce scheduling performance:

            Since reservation scheduling performance consumption is known to grow with the number of pending jobs, use of -R y option is recommended only for those jobs actually queuing for bottleneck resources.

            Source https://stackoverflow.com/questions/42559327

            Community Discussions, Code Snippets contain sources that include Stack Exchange Network

            Vulnerabilities

            Sun GridEngine 5.3 and earlier does not properly check the return value from the OpenSSL EVP_VerifyFinal function, which allows remote attackers to bypass validation of the certificate chain via a malformed SSL/TLS signature for DSA and ECDSA keys, a similar vulnerability to CVE-2008-5077.
            rsh utility in Sun Grid Engine (SGE) before 6.0u7_1 allows local users to gain privileges and execute arbitrary code via unspecified vectors, possibly involving command line arguments.

            Install gridengine

            Get gridengine from [github](https://github.com/hbristow/gridengine) and install using pip:. This will automatically pull and build the dependencies.

            Support

            The gridengine branching schema is set up as follows:.
            Find more information at:

            Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items

            Find more libraries
            CLONE
          • HTTPS

            https://github.com/hbristow/gridengine.git

          • CLI

            gh repo clone hbristow/gridengine

          • sshUrl

            git@github.com:hbristow/gridengine.git

          • Stay Updated

            Subscribe to our newsletter for trending solutions and developer bootcamps

            Agree to Sign up and Terms & Conditions

            Share this Page

            share link

            Explore Related Topics

            Consider Popular Job Scheduling Libraries

            Try Top Libraries by hbristow

            argparse

            by hbristowC++

            cvmatio

            by hbristowC++

            epic

            by hbristowPython

            distance-transform

            by hbristowPython

            eigen

            by hbristowPython