SparkNet | Distributed Neural Networks for Spark
kandi X-RAY | SparkNet Summary
kandi X-RAY | SparkNet Summary
SparkNet is a Scala library typically used in Big Data, Deep Learning, Spark applications. SparkNet has no bugs, it has no vulnerabilities, it has a Permissive License and it has low support. You can download it from GitHub.
Distributed Neural Networks for Spark. Details are available in the paper. Ask questions on the sparknet-users mailing list!.
Distributed Neural Networks for Spark. Details are available in the paper. Ask questions on the sparknet-users mailing list!.
Support
Quality
Security
License
Reuse
Support
SparkNet has a low active ecosystem.
It has 608 star(s) with 178 fork(s). There are 87 watchers for this library.
It had no major release in the last 6 months.
There are 25 open issues and 33 have been closed. On average issues are closed in 141 days. There are 4 open pull requests and 0 closed requests.
It has a neutral sentiment in the developer community.
The latest version of SparkNet is current.
Quality
SparkNet has no bugs reported.
Security
SparkNet has no vulnerabilities reported, and its dependent libraries have no vulnerabilities reported.
License
SparkNet is licensed under the MIT License. This license is Permissive.
Permissive licenses have the least restrictions, and you can use them in most projects.
Reuse
SparkNet releases are not available. You will need to build from source code and install.
Installation instructions, examples and code snippets are available.
Top functions reviewed by kandi - BETA
kandi's functional review helps you automatically verify the functionalities of the libraries and avoid rework.
Currently covering the most popular Java, JavaScript and Python libraries. See a Sample of SparkNet
Currently covering the most popular Java, JavaScript and Python libraries. See a Sample of SparkNet
SparkNet Key Features
No Key Features are available at this moment for SparkNet.
SparkNet Examples and Code Snippets
No Code Snippets are available at this moment for SparkNet.
Community Discussions
Trending Discussions on SparkNet
QUESTION
Spark application stuck in Running state, Initial job has not accepted any resources
Asked 2019-Aug-24 at 12:54
I'm working on a Distributed Deep learning project using Apache Hadoop, Spark and DL4J.
My main issue is when starting my Application on spark it gets to Running state and never gets higher than 10% progress I get this Warning
...ANSWER
Answered 2019-Aug-24 at 12:54Fixed the issue by adding another Slave I don't know why and how it worked but when I added another slave it worked
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install SparkNet
You will probably have to change several fields in this command. For example, the flags --key-pair and --identity-file specify the key pair you will use to connect to the cluster. The flag --slaves specifies the number of Spark workers.
Create an AWS secret key and access key. Instructions here.
Run export AWS_SECRET_ACCESS_KEY= and export AWS_ACCESS_KEY_ID= with the relevant values.
Clone our repository locally.
Start a 5-worker Spark cluster on EC2 by running SparkNet/ec2/spark-ec2 --key-pair=key \ --identity-file=key.pem \ --region=eu-west-1 \ --zone=eu-west-1c \ --instance-type=g2.8xlarge \ --ami=ami-d0833da3 \ --copy-aws-credentials \ --spark-version=1.5.0 \ --spot-price=1.5 \ --no-ganglia \ --user-data SparkNet/ec2/cloud-config.txt \ --slaves=5 \ launch sparknet
SSH to the Spark master as root.
Run bash /root/SparkNet/data/cifar10/get_cifar10.sh to get the Cifar data
Train Cifar on 5 workers using /root/spark/bin/spark-submit --class apps.CifarApp /root/SparkNet/target/scala-2.10/sparknet-assembly-0.1-SNAPSHOT.jar 5
That's all! Information is logged on the master in /root/SparkNet/training_log*.txt.
Obtain the ImageNet data by following the instructions here with wget http://.../ILSVRC2012_img_train.tar wget http://.../ILSVRC2012_img_val.tar This involves creating an account and submitting a request.
On the Spark master, create ~/.aws/credentials with the following content: [default] aws_access_key_id= aws_secret_access_key= and fill in the two fields.
Copy this to the workers with ~/spark-ec2/copy-dir ~/.aws (copy this command exactly because it is somewhat sensitive to the trailing backslashes and that kind of thing).
Create an Amazon S3 bucket with name S3_BUCKET.
Upload the ImageNet data in the appropriate format to S3 with the command python $SPARKNET_HOME/scripts/put_imagenet_on_s3.py $S3_BUCKET \ --train_tar_file=/path/to/ILSVRC2012_img_train.tar \ --val_tar_file=/path/to/ILSVRC2012_img_val.tar \ --new_width=256 \ --new_height=256 This command resizes the images to 256x256, shuffles the training data, and tars the validation files into chunks.
Train ImageNet on 5 workers using /root/spark/bin/spark-submit --class apps.ImageNetApp /root/SparkNet/target/scala-2.10/sparknet-assembly-0.1-SNAPSHOT.jar 5 $S3_BUCKET
We have built the JavaCPP binaries for a couple platforms. They are stored at the following locations:.
Ubuntu with GPUs: http://www.eecs.berkeley.edu/~rkn/snapshot-2016-03-05/
Ubuntu with CPUs: http://www.eecs.berkeley.edu/~rkn/snapshot-2016-03-16-CPU/
CentOS 6 with CPUs: http://www.eecs.berkeley.edu/~rkn/snapshot-2016-03-23-CENTOS6-CPU/
Create an AWS secret key and access key. Instructions here.
Run export AWS_SECRET_ACCESS_KEY= and export AWS_ACCESS_KEY_ID= with the relevant values.
Clone our repository locally.
Start a 5-worker Spark cluster on EC2 by running SparkNet/ec2/spark-ec2 --key-pair=key \ --identity-file=key.pem \ --region=eu-west-1 \ --zone=eu-west-1c \ --instance-type=g2.8xlarge \ --ami=ami-d0833da3 \ --copy-aws-credentials \ --spark-version=1.5.0 \ --spot-price=1.5 \ --no-ganglia \ --user-data SparkNet/ec2/cloud-config.txt \ --slaves=5 \ launch sparknet
SSH to the Spark master as root.
Run bash /root/SparkNet/data/cifar10/get_cifar10.sh to get the Cifar data
Train Cifar on 5 workers using /root/spark/bin/spark-submit --class apps.CifarApp /root/SparkNet/target/scala-2.10/sparknet-assembly-0.1-SNAPSHOT.jar 5
That's all! Information is logged on the master in /root/SparkNet/training_log*.txt.
Obtain the ImageNet data by following the instructions here with wget http://.../ILSVRC2012_img_train.tar wget http://.../ILSVRC2012_img_val.tar This involves creating an account and submitting a request.
On the Spark master, create ~/.aws/credentials with the following content: [default] aws_access_key_id= aws_secret_access_key= and fill in the two fields.
Copy this to the workers with ~/spark-ec2/copy-dir ~/.aws (copy this command exactly because it is somewhat sensitive to the trailing backslashes and that kind of thing).
Create an Amazon S3 bucket with name S3_BUCKET.
Upload the ImageNet data in the appropriate format to S3 with the command python $SPARKNET_HOME/scripts/put_imagenet_on_s3.py $S3_BUCKET \ --train_tar_file=/path/to/ILSVRC2012_img_train.tar \ --val_tar_file=/path/to/ILSVRC2012_img_val.tar \ --new_width=256 \ --new_height=256 This command resizes the images to 256x256, shuffles the training data, and tars the validation files into chunks.
Train ImageNet on 5 workers using /root/spark/bin/spark-submit --class apps.ImageNetApp /root/SparkNet/target/scala-2.10/sparknet-assembly-0.1-SNAPSHOT.jar 5 $S3_BUCKET
We have built the JavaCPP binaries for a couple platforms. They are stored at the following locations:.
Ubuntu with GPUs: http://www.eecs.berkeley.edu/~rkn/snapshot-2016-03-05/
Ubuntu with CPUs: http://www.eecs.berkeley.edu/~rkn/snapshot-2016-03-16-CPU/
CentOS 6 with CPUs: http://www.eecs.berkeley.edu/~rkn/snapshot-2016-03-23-CENTOS6-CPU/
Support
For any new features, suggestions and bugs create an issue on GitHub.
If you have any questions check and ask questions on community page Stack Overflow .
Find more information at:
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page