data-science-r | R code and documentation for Introduction to Data Science
kandi X-RAY | data-science-r Summary
kandi X-RAY | data-science-r Summary
R code and documentation for "Introduction to Data Science" by Jeffrey Stanton.
Support
Quality
Security
License
Reuse
Top functions reviewed by kandi - BETA
Currently covering the most popular Java, JavaScript and Python libraries. See a Sample of data-science-r
data-science-r Key Features
data-science-r Examples and Code Snippets
Community Discussions
Trending Discussions on data-science-r
QUESTION
Can a docker image access hadoop resources? Eg. submit YARN jobs and access HDFS; something like MapR's Datasci. Refinery, but for Hortonworks HDP 3.1. (May assume that the image will be launched on a hadoop cluster node).
Saw the hadoop docs for launching docker applications from hadoop nodes, but was interested in whether could go the "other way" (ie. being able to start a docker image with the conventional docker -ti ...
command and have that application be able to run hadoop jars etc. (assuming that the docker image host is a hadoop node itself)). I understand that MapR hadoop has docker images for doing this, but am interested in using Hortonworks HDP 3.1. Ultimately trying to run h2o hadoop in a docker container.
Anyone know if this is possible or can confirm that this is not possible?
...ANSWER
Answered 2019-Sep-03 at 20:23Yes. As long as you have the client jars and appropriate configs similar to an edge node for your container it should work.
QUESTION
Trying to follow this article (https://mapr.com/blog/how-to-run-data-science-refinery-from-an-edge-node/) for setting up DSR docker image (tag: v1.1_6.0.0_4.1.0_centos7) on an edge node (see here for how to set some of the env.list values: https://mapr.com/docs/61/AdvancedInstallation/Env_Variables_Installer_Container.html). However, once the container is started,
...ANSWER
Answered 2018-Dec-28 at 01:47The DSR image seems to have a bug where, even though the mapr SASL ticket specified in the env.list file exists and is valid, it is not getting copied to the container when it is started. Thus, the container can't connect to the MapR HDFS. To fix this, did...
- Created a file of the same name as the ticket file in the same location on the container as specified in the env.list value specifying where the ticketfile would be
- Manually copy-pasted the contents of the ticketfile from the host to the ticket file we just created in the docker container
- (After waiting a bit (~2 minutes)) Restarted the mapr posix service:
sudo service mapr-posix-client-container restart
After doing this, the container appears to be able to access the HDFS (and submit YARN jobs) fine.
(If anyone has any more information on why this could be happening or if there is a better workaround to get the container working as expected, please let me know).
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install data-science-r
Support
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page