starport | one platform to build , launch and maintain any crypto | Blockchain library
kandi X-RAY | starport Summary
kandi X-RAY | starport Summary
Starport is the all-in-one platform to build, launch, and maintain any crypto application on a sovereign and secured blockchain. It is a developer-friendly interface to the Cosmos SDK, the world's most widely-used blockchain application framework. Starport generates boilerplate code for you, so you can focus on writing business logic.
Support
Quality
Security
License
Reuse
Top functions reviewed by kandi - BETA
Currently covering the most popular Java, JavaScript and Python libraries. See a Sample of starport
starport Key Features
starport Examples and Code Snippets
Community Discussions
Trending Discussions on starport
QUESTION
I'm having s3 endpoint grief. When my instances initialize they can not install docker. Details:
I have ASG instances sitting in a VPC with pub and private subnets. Appropriate routing and EIP/NAT is all stitched up.Instances in private subnets have outbouond 0.0.0.0/0 routed to NAT in respective public subnets. NACLs for public subnet allow internet traffic in and out, the NACLs around private subnets allow traffic from public subnets in and out, traffic out to the internet (and traffic from s3 cidrs in and out). I want it pretty locked down.
- I have DNS and hostnames enabled in my VPC
- I understand NACLs are stateless and have enabled IN and OUTBOUND rules for s3 amazon IP cidr blocks on ephemeral port ranges (yes I have also enabled traffic between pub and private subnets)
- yes I have checked a route was provisioned for my s3 endpoint in my private route tables
- yes I know for sure it is the s3 endpoint causing me grief and not another blunder -> when I delete it and open up my NACLs I can yum update and install docker (as expected) I am not looking for suggestions that require opening up my NACLs, I'm using a VPC gateway endpiont because I want to keep things locked down in the private subnets. I mention this because similar discussions seem to say 'I opened 0.0.0.0/0 on all ports and now x works'
- Should I just bake an AMI with docker installed? That's what I'll do if I can't resolve this. I really wanted to set up my networking so everything is nicely locked down and feel like it should be pretty straight forward utilizing endpoints. Largely this is a networking exercise so I would rather not do this because it avoids solving and understanding the problem.
- I know my other VPC endpoints work perfectly -> Auto-scaling service interface endpoint is performing (I can see it scaling down instances as per the policy), SSM interface endpoint allowing me to use session manager, and ECR endpoint(s) are working in conjunction with s3 gateway endpoint (s3 gateway endpoint is required because image layers are in s3) -> I know this works because if I open up NACLS and delete my s3 endpoint and install docker, then lock everything down again, bring back my s3 gatewayendpoint I can successfully pull my ECR images. SO the s3 gateway endpoint is fine for accessing ecr image layers, but not amazon-linux-extra repos.
- SGs attached to instances are not the problem (instances have default outbound rule)
- I have tried adding increasingly generous policies to my s3 endpoint as I have seen in this 7 year old thread and thought this had to do the trick (yes I subbed my region correctly)
- I strongly feel the solution lies with the s3 gateway policy as discussed in this thread, however have had little luck with my increasingly desperate policies.
Amazon EC2 instance can't update or use yum
another s3 struggle with resolution:
I have tried:
...ANSWER
Answered 2021-Sep-21 at 08:22By the looks of it, you are well aware of what you are trying to achieve. Even though you are saying that it is not the NACLs, I would check them one more time, as sometimes one can easily overlook something minor. Take into account the snippet below taken from this AWS troubleshooting article and make sure that you have the right S3 CIDRs in your rules for the respective region:
Make sure that the network ACLs associated with your EC2 instance's subnet allow the following: Egress on port 80 (HTTP) and 443 (HTTPS) to the Regional S3 service. Ingress on ephemeral TCP ports from the Regional S3 service. Ephemeral ports are 1024-65535. The Regional S3 service is the CIDR for the subnet containing your S3 interface endpoint. Or, if you're using an S3 gateway, the Regional S3 service is the public IP CIDR for the S3 service. Network ACLs don't support prefix lists. To add the S3 CIDR to your network ACL, use 0.0.0.0/0 as the S3 CIDR. You can also add the actual S3 CIDRs into the ACL. However, keep in mind that the S3 CIDRs can change at any time.
Your S3 endpoint policy looks good to me on first look, but you are right that it is very likely that the policy or the endpoint configuration in general could be the cause, so I would re-check it one more time too.
One additional thing that I have observed before is that depending on the AMI you use and your VPC settings (DHCP options set, DNS, etc) sometimes the EC2 instance cannot properly set it's default region in the yum config. Please check whether the files awsregion
and awsdomain
exist within the /etc/yum/vars
directory and what's their content. In your use case, the awsregion should have:
QUESTION
First,I build a cosmos hub testnet for single and local node,just like this
and then I use starport to build a blockchain. the new blockchain can work. But now I find that I have no idea to connect this new blockchain to cosmos hub
like this picture above, I click the button "connect to Cosmos Hub", but it doesn't work. My purpose is to build two blockchains with ibc, and connect these blockchain to the hub. And in this way, the two blockchains can transfer tokens to each other. Anyboby can help me?
...ANSWER
Answered 2021-Aug-08 at 18:25Refer to this video: https://youtu.be/-AuExRijtrA, it neatly explains how can we connect with relayer and send tokens to Cosmos Hub with IBC.
However, if it's not clear from the video, i am explaining the steps below:
You need to remove the relayer configuration if you have used (otherwise you can ignore it)
rm -rf ~/.starport/relayer
Afterwards, start your chain in the development mode using starport c serve -r
Use default values to configure a relayer: starport relayer configure
and send small amount of real ATOMs
to the relayer address on the Hub. The relayer address is printed by the command used above. Wait about 10-15 seconds to make sure the balance has been updated.
starport relayer connect
Wait for the connection to process, now you can send tokens to the Hub:
appd tx ibc-transfer transfer ... ... --from alice
QUESTION
I have created a blockchain with Starport (https://github.com/tendermint/starport).
The command to install I used was starport app github.com/tosch110/stargateapp --sdk-version=stargate
Now trying to run starport serve
and start my blockchain, I get
ANSWER
Answered 2020-Nov-03 at 06:44A relayer is an off-chain program that has a wallet with tokens, watches one or more blockchains for interchain transactions, and sends them to corresponding blockchains.
There is an implementation of a relayer in Go. To install it, run the following commands:
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install starport
Support
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page