spinnaker-for-gcp | Production-ready Spinnaker on GKE | GCP library

 by   GoogleCloudPlatform Shell Version: Current License: No License

kandi X-RAY | spinnaker-for-gcp Summary

kandi X-RAY | spinnaker-for-gcp Summary

spinnaker-for-gcp is a Shell library typically used in Cloud, GCP applications. spinnaker-for-gcp has no bugs, it has no vulnerabilities and it has low support. You can download it from GitHub.

Spinnaker for Google Cloud Platform is a solution for installing and managing Spinnaker on Google Cloud Platform. It consists of an installation and management console, Spinnaker and its microservices, and sample applications.
Support
    Quality
      Security
        License
          Reuse

            kandi-support Support

              spinnaker-for-gcp has a low active ecosystem.
              It has 173 star(s) with 130 fork(s). There are 18 watchers for this library.
              OutlinedDot
              It had no major release in the last 6 months.
              There are 33 open issues and 58 have been closed. On average issues are closed in 17 days. There are 6 open pull requests and 0 closed requests.
              It has a neutral sentiment in the developer community.
              The latest version of spinnaker-for-gcp is current.

            kandi-Quality Quality

              spinnaker-for-gcp has 0 bugs and 0 code smells.

            kandi-Security Security

              spinnaker-for-gcp has no vulnerabilities reported, and its dependent libraries have no vulnerabilities reported.
              spinnaker-for-gcp code analysis shows 0 unresolved vulnerabilities.
              There are 0 security hotspots that need review.

            kandi-License License

              spinnaker-for-gcp does not have a standard license declared.
              Check the repository for any license declaration and review the terms closely.
              OutlinedDot
              Without a license, all rights are reserved, and you cannot use the library in your applications.

            kandi-Reuse Reuse

              spinnaker-for-gcp releases are not available. You will need to build from source code and install.
              Installation instructions, examples and code snippets are available.
              It has 12 lines of code, 2 functions and 1 files.
              It has low code complexity. Code complexity directly impacts maintainability of the code.

            Top functions reviewed by kandi - BETA

            kandi's functional review helps you automatically verify the functionalities of the libraries and avoid rework.
            Currently covering the most popular Java, JavaScript and Python libraries. See a Sample of spinnaker-for-gcp
            Get all kandi verified functions for this library.

            spinnaker-for-gcp Key Features

            No Key Features are available at this moment for spinnaker-for-gcp.

            spinnaker-for-gcp Examples and Code Snippets

            No Code Snippets are available at this moment for spinnaker-for-gcp.

            Community Discussions

            QUESTION

            Submit command line arguments to a pyspark job on airflow
            Asked 2022-Mar-29 at 10:37

            I have a pyspark job available on GCP Dataproc to be triggered on airflow as shown below:

            ...

            ANSWER

            Answered 2022-Mar-28 at 08:18

            You have to pass a Sequence[str]. If you check DataprocSubmitJobOperator you will see that the params job implements a class google.cloud.dataproc_v1.types.Job.

            Source https://stackoverflow.com/questions/71616491

            QUESTION

            Skip first line in import statement using gc.open_by_url from gspread (i.e. add header=0)
            Asked 2022-Mar-16 at 08:12

            What is the equivalent of header=0 in pandas, which recognises the first line as a heading in gspread?

            pandas import statement (correct)

            ...

            ANSWER

            Answered 2022-Mar-16 at 08:12

            Looking at the API documentation, you probably want to use:

            Source https://stackoverflow.com/questions/71418682

            QUESTION

            Automatically Grab Latest Google Cloud Platform Secret Version
            Asked 2022-Mar-01 at 03:01

            I'm trying to grab the latest secret version. Is there a way to do that without specifying the version number? Such as using the keyword "latest". I'm trying to avoid having to iterate through all the secret versions with a for loop as GCP documentation shows:

            ...

            ANSWER

            Answered 2021-Sep-12 at 18:54
            import com.google.cloud.secretmanager.v1.AccessSecretVersionResponse;
            import com.google.cloud.secretmanager.v1.SecretManagerServiceClient;
            import com.google.cloud.secretmanager.v1.SecretVersionName;
            import java.io.IOException;
            
            public class AccessSecretVersion {
            
              public static void accessSecretVersion() throws IOException {
                // TODO(developer): Replace these variables before running the sample.
                String projectId = "your-project-id";
                String secretId = "your-secret-id";
                String versionId = "latest"; //<-- specify version
                accessSecretVersion(projectId, secretId, versionId);
              }
            
              // Access the payload for the given secret version if one exists. The version
              // can be a version number as a string (e.g. "5") or an alias (e.g. "latest").
              public static void accessSecretVersion(String projectId, String secretId, String versionId)
                  throws IOException {
                // Initialize client that will be used to send requests. This client only needs to be created
                // once, and can be reused for multiple requests. After completing all of your requests, call
                // the "close" method on the client to safely clean up any remaining background resources.
                try (SecretManagerServiceClient client = SecretManagerServiceClient.create()) {
                  SecretVersionName secretVersionName = SecretVersionName.of(projectId, secretId, versionId);
            
                  // Access the secret version.
                  AccessSecretVersionResponse response = client.accessSecretVersion(secretVersionName);
            
                  // Print the secret payload.
                  //
                  // WARNING: Do not print the secret in a production environment - this
                  // snippet is showing how to access the secret material.
                  String payload = response.getPayload().getData().toStringUtf8();
                  System.out.printf("Plaintext: %s\n", payload);
                }
              }
            }
            

            Source https://stackoverflow.com/questions/68805240

            QUESTION

            Programmatically Connecting a GitHub repo to a Google Cloud Project
            Asked 2022-Feb-12 at 16:16

            I'm working on a Terraform project that will set up all the GCP resources needed for a large project spanning multiple GitHub repos. My goal is to be able to recreate the cloud infrastructure from scratch completely with Terraform.

            The issue I'm running into is in order to setup build triggers with Terraform within GCP, the GitHub repo that is setting off the trigger first needs to be connected. Currently, I've only been able to do that manually via the Google Cloud Build dashboard. I'm not sure if this is possible via Terraform or with a script but I'm looking for any solution I can automate this with. Once the projects are connected updating everything with Terraform is working fine.

            TLDR; How can I programmatically connect a GitHub project with a GCP project instead of using the dashboard?

            ...

            ANSWER

            Answered 2022-Feb-12 at 16:16

            Currently there is no way to programmatically connect a GitHub repo to a Google Cloud Project. This must be done manually via Google Cloud.

            My workaround is to manually connect an "admin" project, build containers and save them to that project's artifact registry, and then deploy the containers from the registry in the programmatically generated project.

            Source https://stackoverflow.com/questions/69834735

            QUESTION

            Unable to create a new Cloud Function - cloud-client-api-gae
            Asked 2022-Feb-11 at 18:49

            I'm unable to create a Cloud Function in my GCP project using GUI, but have admin roles for GCF, SA and IAM.

            Here is the error message:

            Missing necessary permission iam.serviceAccounts.actAs for cloud-client-api-gae on the service account serviceaccountname@DOMAIN.iam.gserviceaccount.com. Grant the role 'roles/iam.serviceAccountUser' to cloud-client-api-gae on the service account serviceaccountname@DOMAIN.iam.gserviceaccount.com.

            cloud-client-api-gae is not an SA nor User on my IAM list. It must be a creature living underneath Graphical User Interfrace.

            I have Enabled API for GCF, AppEngine and I have Service Account Admin role.

            I had literally 0 search results when googling for cloud-client-api-gae.

            ...

            ANSWER

            Answered 2022-Jan-18 at 13:53

            I contacted GCP support and it seems I was missing a single permission for my user: Service Account User - that's it.

            PS: Person from support didn't know what this thing called "cloud-client-api-gae" is.

            Source https://stackoverflow.com/questions/70756708

            QUESTION

            TypeScript project failing to deploy to App Engine targeting Node 12 or 14, but works with Node 10
            Asked 2022-Jan-16 at 14:32

            I have a TypeScript project that has been deployed several times without any problems to Google App Engine, Standard environment, running Node 10. However, when I try to update the App Engine project to either Node 12 or 14 (by editing the engines.node value in package.json and the runtime value in app.yaml), the deploy fails, printing the following to the console:

            ...

            ANSWER

            Answered 2022-Jan-16 at 14:32

            I encountered the exact same problem and just put typescript in dependencies, not devDependencies.

            It worked after that, but cannot assure that it is due to this change (since I have no proof of that).

            Source https://stackoverflow.com/questions/69153522

            QUESTION

            Dataproc Java client throws NoSuchMethodError setUseJwtAccessWithScope
            Asked 2022-Jan-14 at 19:24

            I am following this article ,for submit a job to an existing Dataproc cluster via a Dataproc API

            For the following line of code :

            ...

            ANSWER

            Answered 2022-Jan-14 at 19:22

            The method com.google.api.gax.core.GoogleCredentialsProvider$Builder com.google.api.gax.core.GoogleCredentialsProvider$Builder.setUseJwtAccessWithScope(boolean) was introduced in com.google.api:gax in version 2.3.0.

            Can you

            1. run mvn dependency:tree and confirm that your version of com.google.api:gax is above version 2.3.0?

            2. upgrade all Google libraries to the latest version?

            Here is a similar issue found on the internet.

            Source https://stackoverflow.com/questions/70640346

            QUESTION

            Apache Beam Cloud Dataflow Streaming Stuck Side Input
            Asked 2022-Jan-12 at 13:12

            I'm currently building PoC Apache Beam pipeline in GCP Dataflow. In this case, I want to create streaming pipeline with main input from PubSub and side input from BigQuery and store processed data back to BigQuery.

            Side pipeline code

            ...

            ANSWER

            Answered 2022-Jan-12 at 13:12

            Here you have a working example:

            Source https://stackoverflow.com/questions/70561769

            QUESTION

            BIG Query command using BAT file
            Asked 2022-Jan-09 at 15:24
            echo Give yearmonth "yyyyMM"
            setlocal enabledelayedexpansion
            SET /p yearmonth= 
            SET ClientName[0]=abc
            SET ClientName[1]=def
            
            SET i = 0
            
            :myLoop
            if defined ClientName[%i%] (
                call bq query --use_legacy_sql=false "CREATE EXTERNAL TABLE `test.!ClientName[%%i]!.%yearmonth%` OPTIONS (format = 'CSV',skip_leading_rows = 1 uris = ['gs://test/!ClientName[%%i]!/AWS/%yearmonth%/Metrics/data/*.csv'])"
                set /a "i+=1"
                GOTO :myLoop
            
            )
            
            ...

            ANSWER

            Answered 2022-Jan-09 at 11:04
            1. It is bad practice to set variables as standalone alphabetical characters like i. One reason is exactly as you have experienced, you have confused for metavariable %%i with a set variable %i%.

            2. You are expanding in the loop, but have not enabledelayedexpansion so there are 2 ways, which we will get to in a second.

            3. setting variables should not have spaces before or after = excluding the likes of set /a

            So, Method 1, without delayedexpansion (note how the variables are used with double %% in the loop with the call command).

            Source https://stackoverflow.com/questions/70640411

            QUESTION

            Vertex AI Model Batch prediction, issue with referencing existing model and input file on Cloud Storage
            Asked 2021-Dec-21 at 14:35

            I'm struggling to correctly set Vertex AI pipeline which does the following:

            1. read data from API and store to GCS and as as input for batch prediction.
            2. get an existing model (Video classification on Vertex AI)
            3. create Batch prediction job with input from point 1.
              As it will be seen, I don't have much experience with Vertex Pipelines/Kubeflow thus I'm asking for help/advice, hope it's just some beginner mistake. this is the gist of the code I'm using as pipeline
            ...

            ANSWER

            Answered 2021-Dec-21 at 14:35

            I'm glad you solved most of your main issues and found a workaround for model declaration.

            For your input.output observation on gcs_source_uris, the reason behind it is because the way the function/class returns the value. If you dig inside the class/methods of google_cloud_pipeline_components you will find that it implements a structure that will allow you to use .outputs from the returned value of the function called.

            If you go to the implementation of one of the components of the pipeline you will find that it returns an output array from convert_method_to_component function. So, in order to have that implemented in your custom class/function your function should return a value which can be called as an attribute. Below is a basic implementation of it.

            Source https://stackoverflow.com/questions/70356856

            Community Discussions, Code Snippets contain sources that include Stack Exchange Network

            Vulnerabilities

            No vulnerabilities reported

            Install spinnaker-for-gcp

            Spinnaker on Google Cloud Platform is a tool for easily installing a production-ready instance of Spinnaker, and for managing that instance over time.
            You access this solution by clicking the Go to Spinnaker for Google Cloud Platform button on the Spinnaker for GCP page in Marketplace. After you've installed Spinnaker for Google Cloud Platform, you can access Spinnaker and the management console from Google Cloud Console. Note: Spinnaker for Google Cloud Platform doesn't support regional clusters. If you intend to install Spinnaker on an existing cluster, it must be zonal. Note: Google recommends that you deploy your resources using an account other than spinnaker-install-account. That account is used to install your spinnaker instance, and resources deployed using that account are installed into the spinnaker namespace by default. This namespace is not indexed, so your deployments will time out before they are deemed stable.
            Start the solution from the Spinnaker for GCP Marketplace page by clicking the Go to Spinnaker for Google Cloud Platform button. When prompted to Open in Cloud Shell, click Proceed. Cloud Shell opens, along with a file tree showing the files in the Spinnaker repository, and instructions. Important: If you've launched the management console at least once before, you might be prompted, in the shell, to resume with the clone you created before, update that clone, or clone a new copy of the repository. The first option is best (cd into the existing directory). Don't clone a new copy. The spinnaker-for-gcp repository is cloned into your Cloud Shell. The flow in the management console guides you through the installation process, presenting you with commands, which you can copy to the Cloud Shell prompt and then execute by pressing Enter. The commands run scripts that automate the process of installing Spinnaker on GKE.
            Start the solution from the Spinnaker for GCP Marketplace page by clicking the Go to Spinnaker for Google Cloud Platform button.
            When prompted to Open in Cloud Shell, click Proceed.
            Follow the instructions shown on the screen.
            In the management console, navigate to step 3, "Scripts for Common Command," and scroll to the bottom of the page. Run the command shown under "Upgrade Management Environment.". The management console is upgraded to include the latest changes.
            In the management console, navigate to step 3, "Scripts for Common Command," and scroll to the bottom of the page.
            Run the command shown under "Upgrade Management Environment."
            You can install your Spinnaker instance or instances on pre-existing infrastructure, instead of having this solution create it new.
            IP aliases enabled, because this uses a hosted Redis instance
            Full Cloud Platform scope for its nodes if you're using the project default service account
            Copy and run the following command (which is also available in step 1 of the installation flow): cloudshell edit \ ~/spinnaker-for-gcp/scripts/install/properties The properties file is opened in the file editor.
            Edit this section of the properties file to identify the Kubernetes cluster on which to install Spinnaker: # If cluster does not exist, it will be created. export GKE_CLUSTER=$DEPLOYMENT_NAME export ZONE=us-west1-b export REGION=us-west1
            Similarly, edit other properties to identify other existing infrastructure and accounts that you want to use, if applicable. For example an existing Cloud Memorystore Redis instance, or a bucket or a service account. In each case, if the infrastructure doesn't exist, the installation script creates it for you.

            Support

            For any new features, suggestions and bugs create an issue on GitHub. If you have any questions check and ask questions on community page Stack Overflow .
            Find more information at:

            Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items

            Find more libraries
            CLONE
          • HTTPS

            https://github.com/GoogleCloudPlatform/spinnaker-for-gcp.git

          • CLI

            gh repo clone GoogleCloudPlatform/spinnaker-for-gcp

          • sshUrl

            git@github.com:GoogleCloudPlatform/spinnaker-for-gcp.git

          • Stay Updated

            Subscribe to our newsletter for trending solutions and developer bootcamps

            Agree to Sign up and Terms & Conditions

            Share this Page

            share link

            Explore Related Topics

            Consider Popular GCP Libraries

            microservices-demo

            by GoogleCloudPlatform

            awesome-kubernetes

            by ramitsurana

            go-cloud

            by google

            infracost

            by infracost

            python-docs-samples

            by GoogleCloudPlatform

            Try Top Libraries by GoogleCloudPlatform

            microservices-demo

            by GoogleCloudPlatformPython

            terraformer

            by GoogleCloudPlatformGo

            training-data-analyst

            by GoogleCloudPlatformJupyter Notebook

            python-docs-samples

            by GoogleCloudPlatformJupyter Notebook

            golang-samples

            by GoogleCloudPlatformGo