dist-keras | Distributed Deep Learning , with a focus
kandi X-RAY | dist-keras Summary
kandi X-RAY | dist-keras Summary
Distributed Keras is a distributed deep learning framework built op top of Apache Spark and Keras, with a focus on "state-of-the-art" distributed optimization algorithms. We designed the framework in such a way that a new distributed optimizer could be implemented with ease, thus enabling a person to focus on research. Several distributed methods are supported, such as, but not restricted to, the training of ensembles and models using data parallel methods. Most of the distributed optimizers we provide, are based on data parallel methods. A data parallel method, as described in [1], is a learning paradigm where multiple replicas of a single model are used to optimize a single objective. Using this approach, we are able to dignificantly reduce the training time of a model. Depending on the parametrization, we also observed that it is possible to achieve better statistical model performance compared to a more traditional approach (e.g., like the SingleTrainer implementation), and yet, spending less wallclock time on the training of the model. However, this is subject to further research. Attention: A rather complete introduction to the problem of Distributed Deep Learning is presented in my Master Thesis Furthermore, the thesis describes includes several novel insights, such as a redefinition of parameter staleness, and several new distributed optimizers such as AGN and ADAG.
Support
Quality
Security
License
Reuse
Top functions reviewed by kandi - BETA
- Train a dataframe
- Allocates a parameter server
- Start the parameter service
- Allocate a worker
- Train the model
- Train a Keras model
- Average a list of models
- Deserialize a keras model
- Allocates a SequentialWorker
- Train a keras model
- Runs the optimizer
- Optimizes the training set
- Handle a pull event
- Transform a dataframe
- Handle a connection
- Normalize a row
- Run the optimizer
- Handle a commit command
- Transform a row into a Spark DataFrame
- Predict predictions for the given iterator
- Reads the data file
- Stops the connection
- Transform row into one - hot encoded row
- Transform a row into a dataframe
- Parse command line arguments
- Run the model
dist-keras Key Features
dist-keras Examples and Code Snippets
Community Discussions
Trending Discussions on dist-keras
QUESTION
Using addPyFiles()
seems to not be adding desiered files to spark job nodes (new to spark so may be missing some basic usage knowledge here).
Attempting to run a script using pyspark and was seeing errors that certain modules are not found for import. Never used spark before, but other posts (from package in question https://github.com/cerndb/dist-keras/issues/36#issuecomment-378918484 and https://stackoverflow.com/a/39779271/8236733) recommended zipping the module and adding to the spark job via sparkContext.addPyFiles(mymodulefiles.zip)
, yet still getting error. The relevant code snippets being...
ANSWER
Answered 2018-Nov-27 at 17:20Fixed problem. Admittedly, solution is not totally spark-related, but leaving question posted for the sake of others who may have similar problem, since the given error message did not make my mistake totally clear from the start.
TLDR: Make sure the package contents (so they should include an __init.py__ in each dir.) of the zip file being loaded in are structured and named the way your code expects.
The package I was trying to load into the spark context via zip was of the form
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install dist-keras
Support
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page