Block Explorer For Bitcoin Testnet Anyblock Analytics
Block Explorer For Bitcoin Testnet Anyblock Analytics
Bitcoin Testnet Faucet
Bitcoin Testnet Faucet
BSV Testnet Explorer
Bitcoin Block Explorer BlockCypher
Scalenet and Testnet4 are online and open for business
Over the years, some people have made use of testnet3 to test out scaling performance, and have spammed testnet3 with 32 MB blocks. This has caused testnet3 to get kinda bloated; the blockchain now takes an hour or so to sync, which slows down development. Other people have wanted to do stress testing, but have specifically wanted to avoid inconveniencing other people by spamming testnet3, and have therefore not done so. This slows down development too. To address this issue, I created two new networks: testnet4 and scalenet. Testnet4 is intended to be a low-volume quick-syncing blockchain which is ideal for testing out new transaction formats or applications. It has a 2 MB default block size limit, and comes with aserti3 parameters that make the difficulty recover quickly to CPU-mineable levels. It should remain easy to sync on a low-end VPS or old laptop. Scalenet is intended to be a high-volume blockchain which is ideal for spamming and stress testing software. It comes with a 256 MB initial default block size limit, and uses aserti3 parameters that make it more suitable for accurately simulating mainnet mining difficulty (though it retains the 20-minute difficulty rule). In order to prevent storage costs from getting unreasonable for a testnet, scalenet will be reset every 6-12 months by invalidating the block at height 10,001 and adding a new checkpoint. Scalenet is intended to be feasible to run on a mid-range desktop computer with some strain. Testnet4 and scalenet are now online and essentially complete. The code for both has been merged into BCHN and Electron Cash. Testnet4 has also been successfully synced to by Knuth, Bitcoin Unlimited, and libbitcoincashj. Block explorers for both are available, thanks to Axel Gembe (who runs the code) and sickpig (who wrote the code): http://tbch4.loping.net:3002/ http://sbch.loping.net:3003/ The testnet4 and scalenet MRs were opened on Aug 19th and Aug 27th, and both were merged on Sep 17th. Scalenet reached height 10,000 on October 3rd. Testnet4 and Scalenet support are present in the master branch of BCHN, and will be included in the next release of BCHN. Some other software (e.g. Electron Cash) already has support in their latest release, but most is still pending. See also: https://bitcoincashresearch.org/t/testnet4-and-scalenet/148/7
Scaling Reddit Community Points with Arbitrum Rollup: a piece of cake
https://preview.redd.it/b80c05tnb9e51.jpg?width=2550&format=pjpg&auto=webp&s=850282c1a3962466ed44f73886dae1c8872d0f31 Submitted for consideration toThe Great Reddit Scaling Bake-Off Baked by the pastry chefs atOffchain Labs Please send questions or comments to [[email protected] ](mailto:[email protected]) 1. Overview We're excited to submit Arbitrum Rollup for consideration to The Great Reddit Scaling Bake-Off. Arbitrum Rollup is the only Ethereum scaling solution that supports arbitrary smart contracts without compromising on Ethereum's security or adding points of centralization. For Reddit, this means that Arbitrum can not only scale the minting and transfer of Community Points, but it can foster a creative ecosystem built around Reddit Community Points enabling points to be used in a wide variety of third party applications. That's right -- you can have your cake and eat it too! Arbitrum Rollup isn't just Ethereum-style. Its Layer 2 transactions are byte-for-byte identical to Ethereum, which means Ethereum users can continue to use their existing addresses and wallets, and Ethereum developers can continue to use their favorite toolchains and development environments out-of-the-box with Arbitrum. Coupling Arbitrum’s tooling-compatibility with its trustless asset interoperability, Reddit not only can scale but can onboard the entire Ethereum community at no cost by giving them the same experience they already know and love (well, certainly know). To benchmark how Arbitrum can scale Reddit Community Points, we launched the Reddit contracts on an Arbitrum Rollup chain. Since Arbitrum provides full Solidity support, we didn't have to rewrite the Reddit contracts or try to mimic their functionality using an unfamiliar paradigm. Nope, none of that. We launched the Reddit contracts unmodified on Arbitrum Rollup complete with support for minting and distributing points. Like every Arbitrum Rollup chain, the chain included a bridge interface in which users can transfer Community Points or any other asset between the L1 and L2 chains. Arbitrum Rollup chains also support dynamic contract loading, which would allow third-party developers to launch custom ecosystem apps that integrate with Community Points on the very same chain that runs the Reddit contracts. 1.1 Why Ethereum Perhaps the most exciting benefit of distributing Community Points using a blockchain is the ability to seamlessly port points to other applications and use them in a wide variety of contexts. Applications may include simple transfers such as a restaurant that allows Redditors to spend points on drinks. Or it may include complex smart contracts -- such as placing Community Points as a wager for a multiparty game or as collateral in a financial contract. The common denominator between all of the fun uses of Reddit points is that it needs a thriving ecosystem of both users and developers, and the Ethereum blockchain is perhaps the only smart contract platform with significant adoption today. While many Layer 1 blockchains boast lower cost or higher throughput than the Ethereum blockchain, more often than not, these attributes mask the reality of little usage, weaker security, or both. Perhaps another platform with significant usage will rise in the future. But today, Ethereum captures the mindshare of the blockchain community, and for Community Points to provide the most utility, the Ethereum blockchain is the natural choice. 1.2 Why Arbitrum While Ethereum's ecosystem is unmatched, the reality is that fees are high and capacity is too low to support the scale of Reddit Community Points. Enter Arbitrum. Arbitrum Rollup provides all of the ecosystem benefits of Ethereum, but with orders of magnitude more capacity and at a fraction of the cost of native Ethereum smart contracts. And most of all, we don't change the experience from users. They continue to use the same wallets, addresses, languages, and tools. Arbitrum Rollup is not the only solution that can scale payments, but it is the only developed solution that can scale both payments and arbitrary smart contracts trustlessly, which means that third party users can build highly scalable add-on apps that can be used without withdrawing money from the Rollup chain. If you believe that Reddit users will want to use their Community Points in smart contracts--and we believe they will--then it makes the most sense to choose a single scaling solution that can support the entire ecosystem, eliminating friction for users. We view being able to run smart contracts in the same scaling solution as fundamentally critical since if there's significant demand in running smart contracts from Reddit's ecosystem, this would be a load on Ethereum and would itself require a scaling solution. Moreover, having different scaling solutions for the minting/distribution/spending of points and for third party apps would be burdensome for users as they'd have to constantly shuffle their Points back and forth. 2. Arbitrum at a glance Arbitrum Rollup has a unique value proposition as it offers a combination of features that no other scaling solution achieves. Here we highlight its core attributes. Decentralized. Arbitrum Rollup is as decentralized as Ethereum. Unlike some other Layer 2 scaling projects, Arbitrum Rollup doesn't have any centralized components or centralized operators who can censor users or delay transactions. Even in non-custodial systems, centralized components provide a risk as the operators are generally incentivized to increase their profit by extracting rent from users often in ways that severely degrade user experience. Even if centralized operators are altruistic, centralized components are subject to hacking, coercion, and potential liability. Massive Scaling. Arbitrum achieves order of magnitude scaling over Ethereum's L1 smart contracts. Our software currently supports 453 transactions-per-second for basic transactions (at 1616 Ethereum gas per tx). We have a lot of room left to optimize (e.g. aggregating signatures), and over the next several months capacity will increase significantly. As described in detail below, Arbitrum can easily support and surpass Reddit's anticipated initial load, and its capacity will continue to improve as Reddit's capacity needs grow. Low cost. The cost of running Arbitrum Rollup is quite low compared to L1 Ethereum and other scaling solutions such as those based on zero-knowledge proofs. Layer 2 fees are low, fixed, and predictable and should not be overly burdensome for Reddit to cover. Nobody needs to use special equipment or high-end machines. Arbitrum requires validators, which is a permissionless role that can be run on any reasonable on-line machine. Although anybody can act as a validator, in order to protect against a “tragedy of the commons” and make sure reputable validators are participating, we support a notion of “invited validators” that are compensated for their costs. In general, users pay (low) fees to cover the invited validators’ costs, but we imagine that Reddit may cover this cost for its users. See more on the costs and validator options below. Ethereum Developer Experience. Not only does Arbitrum support EVM smart contracts, but the developer experience is identical to that of L1 Ethereum contracts and fully compatible with Ethereum tooling. Developers can port existing Solidity apps or write new ones using their favorite and familiar toolchains (e.g. Truffle, Buidler). There are no new languages or coding paradigms to learn. Ethereum wallet compatibility. Just as in Ethereum, Arbitrum users need only hold keys, but do not have to store any coin history or additional data to protect or access their funds. Since Arbitrum transactions are semantically identical to Ethereum L1 transactions, existing Ethereum users can use their existing Ethereum keys with their existing wallet software such as Metamask. Token interoperability. Users can easily transfer their ETH, ERC-20 and ERC-721 tokens between Ethereum and the Arbitrum Rollup chain. As we explain in detail below, it is possible to mint tokens in L2 that can subsequently be withdrawn and recognized by the L1 token contract. Fast finality. Transactions complete with the same finality time as Ethereum L1 (and it's possible to get faster finality guarantees by trading away trust assumptions; see the Arbitrum Rollup whitepaper for details). Non-custodial. Arbitrum Rollup is a non-custodial scaling solution, so users control their funds/points and neither Reddit nor anyone else can ever access or revoke points held by users. Censorship Resistant. Since it's completely decentralized, and the Arbitrum protocol guarantees progress trustlessly, Arbitrum Rollup is just as censorship-proof as Ethereum. Block explorer. The Arbitrum Rollup block explorer allows users to view and analyze transactions on the Rollup chain. Limitations Although this is a bake-off, we're not going to sugar coat anything. Arbitrum Rollup, like any Optimistic Rollup protocol, does have one limitation, and that's the delay on withdrawals. As for the concrete length of the delay, we've done a good deal of internal modeling and have blogged about this as well. Our current modeling suggests a 3-hour delay is sufficient (but as discussed in the linked post there is a tradeoff space between the length of the challenge period and the size of the validators’ deposit). Note that this doesn't mean that the chain is delayed for three hours. Arbitrum Rollup supports pipelining of execution, which means that validators can keep building new states even while previous ones are “in the pipeline” for confirmation. As the challenge delays expire for each update, a new state will be confirmed (read more about this here). So activity and progress on the chain are not delayed by the challenge period. The only thing that's delayed is the consummation of withdrawals. Recall though that any single honest validator knows immediately (at the speed of L1 finality) which state updates are correct and can guarantee that they will eventually be confirmed, so once a valid withdrawal has been requested on-chain, every honest party knows that the withdrawal will definitely happen. There's a natural place here for a liquidity market in which a validator (or someone who trusts a validator) can provide withdrawal loans for a small interest fee. This is a no-risk business for them as they know which withdrawals will be confirmed (and can force their confirmation trustlessly no matter what anyone else does) but are just waiting for on-chain finality. 3. The recipe: How Arbitrum Rollup works For a description of the technical components of Arbitrum Rollup and how they interact to create a highly scalable protocol with a developer experience that is identical to Ethereum, please refer to the following documents: Arbitrum Rollup Whitepaper Arbitrum academic paper (describes a previous version of Arbitrum) 4. Developer docs and APIs For full details about how to set up and interact with an Arbitrum Rollup chain or validator, please refer to our developer docs, which can be found at https://developer.offchainlabs.com/. Note that the Arbitrum version described on that site is older and will soon be replaced by the version we are entering in Reddit Bake-Off, which is still undergoing internal testing before public release. 5. Who are the validators? As with any Layer 2 protocol, advancing the protocol correctly requires at least one validator (sometimes called block producers) that is honest and available. A natural question is: who are the validators? Recall that the validator set for an Arbitrum chain is open and permissionless; anyone can start or stop validating at will. (A useful analogy is to full nodes on an L1 chain.) But we understand that even though anyone can participate, Reddit may want to guarantee that highly reputable nodes are validating their chain. Reddit may choose to validate the chain themselves and/or hire third-party validators.To this end, we have begun building a marketplace for validator-for-hire services so that dapp developers can outsource validation services to reputable nodes with high up-time. We've announced a partnership in which Chainlink nodes will provide Arbitrum validation services, and we expect to announce more partnerships shortly with other blockchain infrastructure providers. Although there is no requirement that validators are paid, Arbitrum’s economic model tracks validators’ costs (e.g. amount of computation and storage) and can charge small fees on user transactions, using a gas-type system, to cover those costs. Alternatively, a single party such as Reddit can agree to cover the costs of invited validators. 6. Reddit Contract Support Since Arbitrum contracts and transactions are byte-for-byte compatible with Ethereum, supporting the Reddit contracts is as simple as launching them on an Arbitrum chain. Minting. Arbitrum Rollup supports hybrid L1/L2 tokens which can be minted in L2 and then withdrawn onto the L1. An L1 contract at address A can make a special call to the EthBridge which deploys a "buddy contract" to the same address A on an Arbitrum chain. Since it's deployed at the same address, users can know that the L2 contract is the authorized "buddy" of the L1 contract on the Arbitrum chain. For minting, the L1 contract is a standard ERC-20 contract which mints and burns tokens when requested by the L2 contract. It is paired with an ERC-20 contract in L2 which mints tokens based on whatever programmer provided minting facility is desired and burns tokens when they are withdrawn from the rollup chain. Given this base infrastructure, Arbitrum can support any smart contract based method for minting tokens in L2, and indeed we directly support Reddit's signature/claim based minting in L2. Batch minting. What's better than a mint cookie? A whole batch! In addition to supporting Reddit’s current minting/claiming scheme, we built a second minting design, which we believe outperforms the signature/claim system in many scenarios. In the current system, Reddit periodically issues signed statements to users, who then take those statements to the blockchain to claim their tokens. An alternative approach would have Reddit directly submit the list of users/amounts to the blockchain and distribute the tokens to the users without the signature/claim process. To optimize the cost efficiency of this approach, we designed an application-specific compression scheme to minimize the size of the batch distribution list. We analyzed the data from Reddit's previous distributions and found that the data is highly compressible since token amounts are small and repeated, and addresses appear multiple times. Our function groups transactions by size, and replaces previously-seen addresses with a shorter index value. We wrote client code to compress the data, wrote a Solidity decompressing function, and integrated that function into Reddit’s contract running on Arbitrum. When we ran the compression function on the previous Reddit distribution data, we found that we could compress batched minting data down to to 11.8 bytes per minting event (averaged over a 6-month trace of Reddit’s historical token grants)compared with roughly 174 bytes of on-chain data needed for the signature claim approach to minting (roughly 43 for an RLP-encoded null transaction + 65 for Reddit's signature + 65 for the user's signature + roughly 8 for the number of Points) . The relative benefit of the two approaches with respect to on-chain call data cost depends on the percentage of users that will actually claim their tokens on chain. With the above figures, batch minting will be cheaper if roughly 5% of users redeem their claims. We stress that our compression scheme is not Arbitrum-specific and would be beneficial in any general-purpose smart contract platform. 8. Benchmarks and costs In this section, we give the full costs of operating the Reddit contracts on an Arbitrum Rollup chain including the L1 gas costs for the Rollup chain, the costs of computation and storage for the L2 validators as well as the capital lockup requirements for staking. Arbitrum Rollup is still on testnet, so we did not run mainnet benchmarks. Instead, we measured the L1 gas cost and L2 workload for Reddit operations on Arbitrum and calculated the total cost assuming current Ethereum gas prices. As noted below in detail, our measurements do not assume that Arbitrum is consuming the entire capacity of Ethereum. We will present the details of our model now, but for full transparency you can also play around with it yourself and adjust the parameters, by copying the spreadsheet found here. Our cost model is based on measurements of Reddit’s contracts, running unmodified (except for the addition of a batch minting function) on Arbitrum Rollup on top of Ethereum. On the distribution of transactions and frequency of assertions. Reddit's instructions specify the following minimum parameters that submissions should support: Over a 5 day period, your scaling PoC should be able to handle:
100,000 point claims (minting & distributing points)
75,000 one-off points burning
We provide the full costs of operating an Arbitrum Rollup chain with this usage under the assumption that tokens are minted or granted to users in batches, but other transactions are uniformly distributed over the 5 day period. Unlike some other submissions, we do not make unrealistic assumptions that all operations can be submitted in enormous batches. We assume that batch minting is done in batches that use only a few percent on an L1 block’s gas, and that other operations come in evenly over time and are submitted in batches, with one batch every five minutes to keep latency reasonable. (Users are probably already waiting for L1 finality, which takes at least that long to achieve.) We note that assuming that there are only 300,000 transactions that arrive uniformly over the 5 day period will make our benchmark numbers lower, but we believe that this will reflect the true cost of running the system. To see why, say that batches are submitted every five minutes (20 L1 blocks) and there's a fixed overhead of c bytes of calldata per batch, the cost of which will get amortized over all transactions executed in that batch. Assume that each individual transaction adds a marginal cost of t. Lastly assume the capacity of the scaling system is high enough that it can support all of Reddit's 300,000 transactions within a single 20-block batch (i.e. that there is more than c + 300,000*t byes of calldata available in 20 blocks). Consider what happens if c, the per-batch overhead, is large (which it is in some systems, but not in Arbitrum). In the scenario that transactions actually arrive at the system's capacity and each batch is full, then c gets amortized over 300,000 transactions. But if we assume that the system is not running at capacity--and only receives 300,000 transactions arriving uniformly over 5 days-- then each 20-block assertion will contain about 200 transactions, and thus each transaction will pay a nontrivial cost due to c. We are aware that other proposals presented scaling numbers assuming that 300,000 transactions arrived at maximum capacity and was executed in a single mega-transaction, but according to our estimates, for at least one such report, this led to a reported gas price that was 2-3 orders of magnitude lower than it would have been assuming uniform arrival. We make more realistic batching assumptions, and we believe Arbitrum compares well when batch sizes are realistic. Our model. Our cost model includes several sources of cost:
L1 gas costs: This is the cost of posting transactions as calldata on the L1 chain, as well as the overhead associated with each batch of transactions, and the L1 cost of settling transactions in the Arbitrum protocol.
Validator’s staking costs: In normal operation, one validator will need to be staked. The stake is assumed to be 0.2% of the total value of the chain (which is assumed to be $1 per user who is eligible to claim points). The cost of staking is the interest that could be earned on the money if it were not staked.
Validator computation and storage: Every validator must do computation to track the chain’s processing of transactions, and must maintain storage to keep track of the contracts’ EVM storage. The cost of computation and storage are estimated based on measurements, with the dollar cost of resources based on Amazon Web Services pricing.
It’s clear from our modeling that the predominant cost is for L1 calldata. This will probably be true for any plausible rollup-based system. Our model also shows that Arbitrum can scale to workloads much larger than Reddit’s nominal workload, without exhausting L1 or L2 resources. The scaling bottleneck will ultimately be calldata on the L1 chain. We believe that cost could be reduced substantially if necessary by clever encoding of data. (In our design any compression / decompression of L2 transaction calldata would be done by client software and L2 programs, never by an L1 contract.) 9. Status of Arbitrum Rollup Arbitrum Rollup is live on Ethereum testnet. All of the code written to date including everything included in the Reddit demo is open source and permissively licensed under the Apache V2 license. The first testnet version of Arbitrum Rollup was released on testnet in February. Our current internal version, which we used to benchmark the Reddit contracts, will be released soon and will be a major upgrade. Both the Arbitrum design as well as the implementation are heavily audited by independent third parties. The Arbitrum academic paper was published at USENIX Security, a top-tier peer-reviewed academic venue. For the Arbitrum software, we have engaged Trail of Bits for a security audit, which is currently ongoing, and we are committed to have a clean report before launching on Ethereum mainnet. 10. Reddit Universe Arbitrum Rollup Chain The benchmarks described in this document were all measured using the latest internal build of our software. When we release the new software upgrade publicly we will launch a Reddit Universe Arbitrum Rollup chain as a public demo, which will contain the Reddit contracts as well as a Uniswap instance and a Connext Hub, demonstrating how Community Points can be integrated into third party apps. We will also allow members of the public to dynamically launch ecosystem contracts. We at Offchain Labs will cover the validating costs for the Reddit Universe public demo. If the folks at Reddit would like to evaluate our software prior to our public demo, please email us at [email protected] and we'd be more than happy to provide early access. 11. Even more scaling: Arbitrum Sidechains Rollups are an excellent approach to scaling, and we are excited about Arbitrum Rollup which far surpasses Reddit's scaling needs. But looking forward to Reddit's eventual goal of supporting hundreds of millions of users, there will likely come a time when Reddit needs more scaling than any Rollup protocol can provide. While Rollups greatly reduce costs, they don't break the linear barrier. That is, all transactions have an on-chain footprint (because all calldata must be posted on-chain), albeit a far smaller one than on native Ethereum, and the L1 limitations end up being the bottleneck for capacity and cost. Since Ethereum has limited capacity, this linear use of on-chain resources means that costs will eventually increase superlinearly with traffic. The good news is that we at Offchain Labs have a solution in our roadmap that can satisfy this extreme-scaling setting as well: Arbitrum AnyTrust Sidechains. Arbitrum Sidechains are similar to Arbitrum Rollup, but deviate in that they name a permissioned set of validators. When a chain’s validators agree off-chain, they can greatly reduce the on-chain footprint of the protocol and require almost no data to be put on-chain. When validators can't reach unanimous agreement off-chain, the protocol reverts to Arbitrum Rollup. Technically, Arbitrum Sidechains can be viewed as a hybrid between state channels and Rollup, switching back and forth as necessary, and combining the performance and cost that state channels can achieve in the optimistic case, with the robustness of Rollup in other cases. The core technical challenge is how to switch seamlessly between modes and how to guarantee that security is maintained throughout. Arbitrum Sidechains break through this linear barrier, while still maintaining a high level of security and decentralization. Arbitrum Sidechains provide the AnyTrust guarantee, which says that as long as any one validator is honest and available (even if you don't know which one will be), the L2 chain is guaranteed to execute correctly according to its code and guaranteed to make progress. Unlike in a state channel, offchain progress does not require unanimous consent, and liveness is preserved as long as there is a single honest validator. Note that the trust model for Arbitrum Sidechains is much stronger than for typical BFT-style chains which introduce a consensus "voting" protocols among a small permissioned group of validators. BFT-based protocols require a supermajority (more than 2/3) of validators to agree. In Arbitrum Sidechains, by contrast, all you need is a single honest validator to achieve guaranteed correctness and progress. Notice that in Arbitrum adding validators strictly increases security since the AnyTrust guarantee provides correctness as long as any one validator is honest and available. By contrast, in BFT-style protocols, adding nodes can be dangerous as a coalition of dishonest nodes can break the protocol. Like Arbitrum Rollup, the developer and user experiences for Arbitrum Sidechains will be identical to that of Ethereum. Reddit would be able to choose a large and diverse set of validators, and all that they would need to guarantee to break through the scaling barrier is that a single one of them will remain honest. We hope to have Arbitrum Sidechains in production in early 2021, and thus when Reddit reaches the scale that surpasses the capacity of Rollups, Arbitrum Sidechains will be waiting and ready to help. While the idea to switch between channels and Rollup to get the best of both worlds is conceptually simple, getting the details right and making sure that the switch does not introduce any attack vectors is highly non-trivial and has been the subject of years of our research (indeed, we were working on this design for years before the term Rollup was even coined). 12. How Arbitrum compares We include a comparison to several other categories as well as specific projects when appropriate. and explain why we believe that Arbitrum is best suited for Reddit's purposes. We focus our attention on other Ethereum projects. Payment only Rollups. Compared to Arbitrum Rollup, ZK-Rollups and other Rollups that only support token transfers have several disadvantages:
As outlined throughout the proposal, we believe that the entire draw of Ethereum is in its rich smart contracts support which is simply not achievable with today's zero-knowledge proof technology. Indeed, scaling with a ZK-Rollup will add friction to the deployment of smart contracts that interact with Community Points as users will have to withdraw their coins from the ZK-Rollup and transfer them to a smart contract system (like Arbitrum). The community will be best served if Reddit builds on a platform that has built-in, frictionless smart-contract support.
All other Rollup protocols of which we are aware employ a centralized operator. While it's true that users retain custody of their coins, the centralized operator can often profit from censoring, reordering, or delaying transactions. A common misconception is that since they're non-custodial protocols, a centralized sequencer does not pose a risk but this is incorrect as the sequencer can wreak havoc or shake down users for side payments without directly stealing funds.
Sidechain type protocols can eliminate some of these issues, but they are not trustless. Instead, they require trust in some quorum of a committee, often requiring two-third of the committee to be honest, compared to rollup protocols like Arbitrum that require only a single honest party. In addition, not all sidechain type protocols have committees that are diverse, or even non-centralized, in practice.
Plasma-style protocols have a centralized operator and do not support general smart contracts.
13. Concluding Remarks While it's ultimately up to the judges’ palate, we believe that Arbitrum Rollup is the bakeoff choice that Reddit kneads. We far surpass Reddit's specified workload requirement at present, have much room to optimize Arbitrum Rollup in the near term, and have a clear path to get Reddit to hundreds of millions of users. Furthermore, we are the only project that gives developers and users the identical interface as the Ethereum blockchain and is fully interoperable and tooling-compatible, and we do this all without any new trust assumptions or centralized components. But no matter how the cookie crumbles, we're glad to have participated in this bake-off and we thank you for your consideration. About Offchain Labs Offchain Labs, Inc. is a venture-funded New York company that spun out of Princeton University research, and is building the Arbitrum platform to usher in the next generation of scalable, interoperable, and compatible smart contracts. Offchain Labs is backed by Pantera Capital, Compound VC, Coinbase Ventures, and others. Leadership Team Ed Felten Ed Felten is Co-founder and Chief Scientist at Offchain Labs. He is on leave from Princeton University, where he is the Robert E. Kahn Professor of Computer Science and Public Affairs. From 2015 to 2017 he served at the White House as Deputy United States Chief Technology Officer and senior advisor to the President. He is an ACM Fellow and member of the National Academy of Engineering. Outside of work, he is an avid runner, cook, and L.A. Dodgers fan. Steven Goldfeder Steven Goldfeder is Co-founder and Chief Executive Officer at Offchain Labs. He holds a PhD from Princeton University, where he worked at the intersection of cryptography and cryptocurrencies including threshold cryptography, zero-knowledge proof systems, and post-quantum signatures. He is a co-author of Bitcoin and Cryptocurrency Technologies, the leading textbook on cryptocurrencies, and he has previously worked at Google and Microsoft Research, where he co-invented the Picnic signature algorithm. When not working, you can find Steven spending time with his family, taking a nature walk, or twisting balloons. Harry Kalodner Harry Kalodner is Co-founder and Chief Technology Officer at Offchain Labs where he leads the engineering team. Before the company he attended Princeton as a Ph.D candidate where his research explored economics, anonymity, and incentive compatibility of cryptocurrencies, and he also has worked at Apple. When not up at 3:00am writing code, Harry occasionally sleeps.
Syscoin Platform’s Great Reddit Scaling Bake-off Proposal
https://preview.redd.it/rqt2dldyg8e51.jpg?width=1044&format=pjpg&auto=webp&s=777ae9d4fbbb54c3540682b72700fc4ba3de0a44 We are excited to participate and present Syscoin Platform's ideal characteristics and capabilities towards a well-rounded Reddit Community Points solution! Our scaling solution for Reddit Community Points involves 2-way peg interoperability with Ethereum. This will provide a scalable token layer built specifically for speed and high volumes of simple value transfers at a very low cost, while providing sovereign ownership and onchain finality. Token transfers scale by taking advantage of a globally sorting mempool that provides for probabilistically secure assumptions of “as good as settled”. The opportunity here for token receivers is to have an app-layer interactivity on the speed/security tradeoff (99.9999% assurance within 10 seconds). We call this Z-DAG, and it achieves high-throughput across a mesh network topology presently composed of about 2,000 geographically dispersed full-nodes. Similar to Bitcoin, however, these nodes are incentivized to run full-nodes for the benefit of network security, through a bonded validator scheme. These nodes do not participate in the consensus of transactions or block validation any differently than other nodes and therefore do not degrade the security model of Bitcoin’s validate first then trust, across every node. Each token transfer settles on-chain. The protocol follows Bitcoin core policies so it has adequate code coverage and protocol hardening to be qualified as production quality software. It shares a significant portion of Bitcoin’s own hashpower through merged-mining. This platform as a whole can serve token microtransactions, larger settlements, and store-of-value in an ideal fashion, providing probabilistic scalability whilst remaining decentralized according to Bitcoin design. It is accessible to ERC-20 via a permissionless and trust-minimized bridge that works in both directions. The bridge and token platform are currently available on the Syscoin mainnet. This has been gaining recent attention for use by loyalty point programs and stablecoins such as Binance USD.
Syscoin Foundation identified a few paths for Reddit to leverage this infrastructure, each with trade-offs. The first provides the most cost-savings and scaling benefits at some sacrifice of token autonomy. The second offers more preservation of autonomy with a more narrow scope of cost savings than the first option, but savings even so. The third introduces more complexity than the previous two yet provides the most overall benefits. We consider the third as most viable as it enables Reddit to benefit even while retaining existing smart contract functionality. We will focus on the third option, and include the first two for good measure.
Distribution, burns and user-to-user transfers of Reddit Points are entirely carried out on the Syscoin network. This full-on approach to utilizing the Syscoin network provides the most scalability and transaction cost benefits of these scenarios. The tradeoff here is distribution and subscription handling likely migrating away from smart contracts into the application layer.
The Reddit Community Points ecosystem can continue to use existing smart contracts as they are used today on the Ethereum mainchain. Users migrate a portion of their tokens to Syscoin, the scaling network, to gain much lower fees, scalability, and a proven base layer, without sacrificing sovereign ownership. They would use Syscoin for user-to-user transfers. Tips redeemable in ten seconds or less, a high-throughput relay network, and onchain settlement at a block target of 60 seconds.
Integration between Matic Network and Syscoin Platform - similar to Syscoin’s current integration with Ethereum - will provide Reddit Community Points with EVM scalability (including the Memberships ERC777 operator) on the Matic side, and performant simple value transfers, robust decentralized security, and sovereign store-of-value on the Syscoin side. It’s “the best of both worlds”. The trade-off is more complex interoperability.
Syscoin + Matic Integration
Matic and Blockchain Foundry Inc, the public company formed by the founders of Syscoin, recently entered a partnership for joint research and business development initiatives. This is ideal for all parties as Matic Network and Syscoin Platform provide complementary utility. Syscoin offers characteristics for sovereign ownership and security based on Bitcoin’s time-tested model, and shares a significant portion of Bitcoin’s own hashpower. Syscoin’s focus is on secure and scalable simple value transfers, trust-minimized interoperability, and opt-in regulatory compliance for tokenized assets rather than scalability for smart contract execution. On the other hand, Matic Network can provide scalable EVM for smart contract execution. Reddit Community Points can benefit from both. Syscoin + Matic integration is actively being explored by both teams, as it is helpful to Reddit, Ethereum, and the industry as a whole.
Total cost for these 100k transactions: $0.63 USD See the live fee comparison for savings estimation between transactions on Ethereum and Syscoin. Below is a snapshot at time of writing: ETH price: $318.55 ETH gas price: 55.00 Gwei ($0.37) Syscoin price: $0.11 Snapshot of live fee comparison chart Z-DAG provides a more efficient fee-market. A typical Z-DAG transaction costs 0.0000582 SYS. Tokens can be safely redeemed/re-spent within seconds or allowed to settle on-chain beforehand. The costs should remain about this low for microtransactions. Syscoin will achieve further reduction of fees and even greater scalability with offchain payment channels for assets, with Z-DAG as a resilience fallback. New payment channel technology is one of the topics under research by the Syscoin development team with our academic partners at TU Delft. In line with the calculation in the Lightning Networks white paper, payment channels using assets with Syscoin Core will bring theoretical capacity for each person on Earth (7.8 billion) to have five on-chain transactions per year, per person, without requiring anyone to enter a fee market (aka “wait for a block”). This exceeds the minimum LN expectation of two transactions per person, per year; one to exist on-chain and one to settle aggregated value.
Tools to simplify using Syscoin Bridge as a service with dapps and wallets will be released some time after implementation of Syscoin Core 4.2. These will be based upon the same processes which are automated in the current live Sysethereum Dapp that is functioning with the Syscoin mainnet.
The Syscoin Ethereum Bridge is secured by Agent nodes participating in a decentralized and incentivized model that involves roles of Superblock challengers and submitters. This model is open to participation. The benefits here are trust-minimization, permissionless-ness, and potentially less legal/regulatory red-tape than interop mechanisms that involve liquidity providers and/or trading mechanisms. The trade-off is that due to the decentralized nature there are cross-chain settlement times of one hour to cross from Ethereum to Syscoin, and three hours to cross from Syscoin to Ethereum. We are exploring ways to reduce this time while maintaining decentralization via zkp. Even so, an “instant bridge” experience could be provided by means of a third-party liquidity mechanism. That option exists but is not required for bridge functionality today. Typically bridges are used with batch value, not with high frequencies of smaller values, and generally it is advantageous to keep some value on both chains for maximum availability of utility. Even so, the cross-chain settlement time is good to mention here.
Ethereum -> Syscoin: Matic or Ethereum transaction fee for bridge contract interaction, negligible Syscoin transaction fee for minting tokens Syscoin -> Ethereum: Negligible Syscoin transaction fee for burning tokens, 0.01% transaction fee paid to Bridge Agent in the form of the ERC-20, Matic or Ethereum transaction fee for contract interaction.
Zero-Confirmation Directed Acyclic Graph is an instant settlement protocol that is used as a complementary system to proof-of-work (PoW) in the confirmation of Syscoin service transactions. In essence, a Z-DAG is simply a directed acyclic graph (DAG) where validating nodes verify the sequential ordering of transactions that are received in their memory pools. Z-DAG is used by the validating nodes across the network to ensure that there is absolute consensus on the ordering of transactions and no balances are overflowed (no double-spends).
Unique fee-market that is more efficient for microtransaction redemption and settlement
Uses decentralized means to enable tokens with value transfer scalability that is comparable or exceeds that of credit card networks
Provides high throughput and secure fulfillment even if blocks are full
Probabilistic and interactive
99.9999% security assurance within 10 seconds
Can serve payment channels as a resilience fallback that is faster and lower-cost than falling-back directly to a blockchain
Each Z-DAG transaction also settles onchain through Syscoin Core at 60-second block target using SHA-256 Proof of Work consensus
Z-DAG enables the ideal speed/security tradeoff to be determined per use-case in the application layer. It minimizes the sacrifice required to accept and redeem fast transfers/payments while providing more-than-ample security for microtransactions. This is supported on the premise that a Reddit user receiving points does need security yet generally doesn’t want nor need to wait for the same level of security as a nation-state settling an international trade debt. In any case, each Z-DAG transaction settles onchain at a block target of 60 seconds.
Syscoin 3.0 White Paper (4.0 white paper is pending. For improved scalability and less blockchain bloat, some features of v3 no longer exist in current v4: Specifically Marketplace Offers, Aliases, Escrow, Certificates, Pruning, Encrypted Messaging)
16MB block bandwidth per minute assuming segwit witness carrying transactions, and transactions ~200 bytes on average
SHA256 merge mined with Bitcoin
UTXO asset layer, with base Syscoin layer sharing identical security policies as Bitcoin Core
Z-DAG on asset layer, bridge to Ethereum on asset layer
On-chain scaling with prospect of enabling enterprise grade reliable trustless payment processing with on/offchain hybrid solution
Focus only on Simple Value Transfers. MVP of blockchain consensus footprint is balances and ownership of them. Everything else can reduce data availability in exchange for scale (Ethereum 2.0 model). We leave that to other designs, we focus on transfers.
Future integrations of MAST/Taproot to get more complex value transfers without trading off trustlessness or decentralization.
Zero-knowledge Proofs are a cryptographic new frontier. We are dabbling here to generalize the concept of bridging and also verify the state of a chain efficiently. We also apply it in our Digital Identity projects at Blockchain Foundry (a publicly traded company which develops Syscoin softwares for clients). We are also looking to integrate privacy preserving payment channels for off-chain payments through zkSNARK hub & spoke design which does not suffer from the HTLC attack vectors evident on LN. Much of the issues plaguing Lightning Network can be resolved using a zkSNARK design whilst also providing the ability to do a multi-asset payment channel system. Currently we found a showstopper attack (American Call Option) on LN if we were to use multiple-assets. This would not exist in a system such as this.
Web3 and mobile wallets are under active development by Blockchain Foundry Inc as WebAssembly applications and expected for release not long after mainnet deployment of Syscoin Core 4.2. Both of these will be multi-coin wallets that support Syscoin, SPTs, Ethereum, and ERC-20 tokens. The Web3 wallet will provide functionality similar to Metamask. Syscoin Platform and tokens are already integrated with Blockbook. Custom hardware wallet support currently exists via ElectrumSys. First-class HW wallet integration through apps such as Ledger Live will exist after 4.2. Current supported wallets Syscoin Spark Desktop Syscoin-Qt
AMA Recap telos Foundation with Crypto Hunters On August 02, 2020 at 12:00 WIB Indonesia Time / August 01 2020 at 10:00 PM ( PST ) in the Crypto Hunter Telegram Group, AMA TELOS started with Mr.Douglas as guest speaker and Gus Fahlev from Crypto Hunters as moderator. When campaigning, 10 lucky AMA participants when asking questions on Google forms and AMA sessions will get a total TELOS ( TLOS ) prize of $100. The following is a summary of AMA questions and answers announced by the moderator and Segment 1 Question 1: Can you explain us, what is Telos? Answer: Telos is a blockchain platform for smart contracts. It is a low latencya new block every half second, high capacitycurrently in the top 2 blockchains in transactions per day, according to Blocktivity.info, and no transaction fee blockchain. Telos also has many unique features that allow developers to make better, dapps, such as our Telos Decide governance engine. Question 2: what ecosystem is used by telos? Answer: Telos is its own Layer-1 blockchain, not a token on another blockchain. The technology behind Telos is EOSIO, the same technology used by EOS and WAX, for example. Question 3: I see that Telos uses EOSIO platform, what are the very significant advantages that distinguish Telos from other projects? Answer: Telos uses the EOSIO platform but we have built several additional tools. Some of these add more security and resiliency to the blockchain, such as testing block producers and removing non-performant ones, but most are related to development. Telos provides attractive development tools that arent available elsewhere. Telos Decide is a governance platform that lets any group create self-governance tools easily. These run on Telos at very little cost and can provide all kinds of voting, elections, initiative ballots, committee management and funds allocation. Telos also has Telos EVM, an Ethereum virtual machine that can run Ethereum Solidity contracts at hundreds of times the speed of Ethereum and with no costs. Another Telos technology that is deploying soon is dStor, which is a decentralized cloud storage system associated with Telos so that dapps can store files controlled by blockchain contracts. Question 4: At what stage is Teloa Road Map now? what are the latest updates currently being realized? Answer: Telos launched its mainnet in December 2018 and has so far produced over 100,000,000 blocks without ever stopping or rolling back the chain. This is likely a record for a public blockchain. We have an ongoing group Telos Core Developers who build and maintain the code and are paid by our Telos Works funding system that is voted by the Telos token holders. Telos is a leader in blockchain governance and regularly amends its governance rules based on smart contract powered voting called Telos Amend. You can see the current Telos governance rules as stored live on the blockchain at tbnoa.org. The most recent updates were adding new features to Telos Decide to make it more powerful, implementing EOSIO v2.0 which increased the capacity of Telos about 8-10 times what it previously was, and implementing Telos EVM on our Testnet. We are currently working on better interfaces for Telos Decide voting, and building more infrastructure around Telos EVM so that it is ready to deploy on our mainnet. Question 5: Is telos currently available on an exchange? and is it ready to be traded? Answer: Telos has been trading on exchanges for over a year. The largest exchanges are Probit, CoinTiger, CoinLim, and P2PB2B. Other exchanges include Newdex and Alcor. We expect to be listed on larger exchanges in the near future. Question 6: Now is the time when defi tokens begin to develop, can telos be categorized as a defi project? and what strategies for this year and in the years to come prepared by telos? Answer: Telos is a smart contract platform, but it already has many DeFi tools built for it including REX staking rewards with a current yield of ~19% APR, smart contract controlled token swaps (like Bancor) with no counterparty called Telos Swaps, a common liquidity pool/order book shared by multiple DEXs to improve liquidity called EvolutionDEX. Wrapped BTC, ETH, XRP, EOS, and other tokens can be brought to Telos and exchanged or used via smart contracts through Transledger. We have more DeFi tools coming all the time including two new offerings in the next few weeks that will be the first of their kind. Question 7: Governance is an important topic in blockchain and Telos is considered a leader in this area. Why is that? Answer: Telos is among the top blockchain projects in terms of how it empowers its users to guide the growth of the chainalong the likes of Tezos or new DeFi tokens that offer governance coins. Telos users continuously elect the validating nodes, called Block Producers, that operate the network based on a set of governance documents such as the Telos Blockchain Network Operating Agreement (TBNOA). These are all stored entirely on-chain (viewable at tbnoa.org) and can be modified by smart contract through blockchain voting using Telos Amend. You can see examples of this at https://chainspector.io/governance/ratify-proposals Telos also has a robust user-voted funding mechanism called Telos Works that has funded many projects and is one of the more successful blockchain incubators around. Voting for all of these can be done in a number of ways including block explorers, wallets like Sqrl (desktop) and Telos Wallet (mobile), telos.net and Chainspector (https://chainspector.io/governance/telos-works). But Telos goes beyond any other chain-level governance by making all of these features and more available to any dapp on Telos through Telos Decide governance engine, making it easy for any dapp or DAO to add robust, highly customized voting. Segment 2 from google form Question 1: Defi projects are now trending whether telos will also go to Defi projects, to increase investors or the community? Answer: Yes, we have several DeFi tools on Telos that can work together: Telos Swaps is an automated, zero-counterparty token swapping smart contract where you can exchange any Telos tokens you may want at any time. Telos has DEXs and uses a common order book called EvolutionDEX that's available to any DEX so that a buy order on one can be matched against a sell order on another. This greatly increases liquidity for traders. We have staking rewards though the Resource EXchange (REX) with rewards currently at about 19% APR. We also have "wrapped" BTC, ETH, and other tokens that can be traded on Telos or used by its smart contracts at half-second transaction times with no transaction fees. This makes Telos a Bitcoin or Ethereum second layer or state channel that's much faster even than Lightning Network and has no fees once the BTC has been brought to Telos. Question 2: Telos aim is to build a new global economy could you explain how whole ecosystem works? There are already many centralized competitors so what is decentralization aspect in telos? Answer: Telos is one of the most decentralized blockchain's in the world. It is operated by 51 validators (block producers) who validate blocks in any month. These are voted for on an ongoing basis by Telos account holders. Telos is also economically decentralized with no large whales like Bitcoin, Ethereum, XRP or EOS because Telos never performed an ICO and limited the size of genesis accounts to 40,000 TLOS max. Telos is also geographically decentralized with users and block producers on every continent but Antarctica and in numerous countries. The is a large amount in North America and Western Europe, but also in Asia, Australia, and large contingents in Latin America and Africa. Telos has had a Block Producer in Indonesia since the beginning and some dapps on Telos are based in Indonesia as well, like SEEDS, for example. Question 3: Most investors focus only on the token price in the short term instead of the real value of the project. Can #TELOStell me the benefits for investors holding #TELOSthe long term? Answer: That's true about crypto speculators and traders, certainly. Traders are usually looking for coins with good positive momentuum that they hope will continue. But these are often pump and dumps where a few people get in early, pump the price, and then get out at the expense of new investors. That's very unfortunate. Telos isn't like this. One reason is that there aren't large whales who can easily manipulate the price. Telos seems to be greatly undervalued compared to its peers. Telos has capacity like EOS and well above XRP, XML, Tron, Ethereum. But its value is miniscule relative to these. Telos is a leader in blockchain governance like Tezos but its marketcap is tiny in comparison. Telos onboarded 100,000 new accounts last month and is appearing in the leading crypto press every week with new dapps or developments. So there's some disconnect between the value of Telos and the price. In my experience, these tend to equalize once more people learn about a project. Question 4: Eos Problems and How Telos Will Solve Them? Answer: Telos originally set out to solve problems with EOS. It was successful in this and now Telos stands on it's own and our roadmap is more about empowering users. In short, these are some of the EOS problems we already solved: RAM speculation - Telos had a plan to reduce RAM speculation through a published guidance price that has been extremely successful. The RAM price is guided by market forces but has remained within 10% of the guidance price since launch. CPU resources - Telos implemented the Telos Resource Improved Management Plan many months ago which was a 7-point approach to making EIDOS-type resource mining unprofitable on Telos. It has largely been successful and Telos has not experienced any resource shortages. Exchange Collusion/Voting - Telos governance does not permit Exchanges to vote with user tokens. This prevent voting situations seen on EOS or STEEM. Block Producer collusion - Telos has minimum requirements for block producers and do not allow anyone to own more than one block producer. Those who are found doing so (there have been about 3 cases so far) have been removed and sanctioned in accordance with the rules of the TBNOA. Question 5: What ecosystems do telos use? and why telos prefers to use EOS network over BEP2 or ERC20? what layer is used telos, can you please explain? Answer: uses the EOSIO protocol because it is the fastest and most powerful in the world and it also receives the fastest upgrades and ongoing development compared to other blockchain technologies. EOS and WAX also use the EOSIO protocol but they are completely different chains. Telos is a Layer 1 protocol, meaning that it is its own blockchain that other dapps and smart contracts deploy upon. One thing that happens when a blockchain like Telos has much, much higher speed and capacity than others like Bitcoin or Ethereum is that Telos can actually run those other blockchains better on its own platform than they can natively. For example, a number of tokens can come in to Telos as wrapped tokens. BTC, ETH, XRP are all current examples of tokens that can be on Telos as wrapped tokens. Once there, these can all be moved around with half-second transaction times and no transaction fees, so they are a better second layer for Bitcoin or Ethereum than Lightning Network or Loom. Telos can also emulate other chains, which we are doing using Telos EVM which emulates the Ethereum Virtual Machine at about 300 times faster and with no gas fees or congestion compared to Ethereum native deployment. Telos can run Ethereum (Solidity) smart contracts without any changes required. Telos EVM is already deployed on the Telos Testnet and will move to our mainnet soon. So anyone who wants to run ERC-20 tokens on Telos can do so easily and they will be faster and with much less cost than running the same contract on Ethereum. Segment 3 free asking Question: I am happy to see new things created by the Telos team. Like What concept did you build in 2020 to make Telos superior? Answer: Currently, I think Telos Decide is the most unique and powerful feature we have built. There are all kinds of organizations that need to vote. Apartment buildings, school boards, unions, tribes, youth sports leagues, city councils. Voting is hard, time consuming, and expensive for many. Telos Decide makes voting easy, convenient, and transparent. That will be a major improvement and disrupt old style voting. It also goes for buisnesses and corporate governance. Even before COVID it was important, but now people can't really gather in one place so fraud-proof voting is very important. No one has the tools that Telos has. And if they try to copy us, well, we are already way out ahead working on the next features. Question: If we look about partnerships, Telos has many partnership ! so what's the importance of that partnership for Telos? And How will you protect the value of Telos to your partners or investors ?? Answer: Many of the partnerships are dapps that have decided to deploy on Telos and receive some level of help from the TCD or Telos Foundation to do so. Once a dapp deploys on a chain, it really is like a long term partnership. Many dapps will become block producers as well and join in the governance of Telos. I suspect that in a few years, most block producers will be the large dapps on the platform with just a few remaining like my company GoodBlock. Of course, we will have our own apps out as well so I guess we'll be developers too. Telos is very fiscally responsible for investors. We spend little. There has not been any actual inflation on the chain in almost a year. (The token supply has remained unchanged at about 355M TLOS) we are actively working with dapps to bring more to Telos and exchanges and other services like fiat on- and off-ramps to increase value for users. Question: In challenging crypto market condition any project is really difficult to survive and we are witnessing that there are many platforms . What is telos project plan for surviving in this long blockchain marathon? In this plan, what motivates long term investors and believers? Answer: True. While we currently have a low token price, Telos as a DPOS chain can be maintained and grow without a massive army of miners and still maintain BFT. But the risk is really not whether Telos can continue. Already there are enough dapps that if the block producers went away somehow (not gonna happen) the dapps would just run the chain themselves. But with 100,000 new users last month and new dapps all the time, we are looking to join the top 5 dapp platforms on DappRadar soon. Survival as a project is not in question. One of the big reasons is that we never did any ICO and Telos is not a company. So regulatory risks aren't there and there's no company to go bankrupt or fail. We have already developed a bootstrapped system to pay block producers and core developers. So we aren't like a company that will run out of runway sometime. Question: Could you explain what is DSTOR? What will it contribute to your ecosystem? Answer: dStor is a decentralized cloud storage system that will have the performance of AWS or Azure with much lower costs and true decentralization. It's based on a highly modified version of IPFS that we have applied for patents for our implementation. It means that dapps will be able to store data like files, images, sound, etc. in a decentralized way. Question: Trust and security is very important in any business , what makes investors , customer and users safe secure when working with TELOS?? Answer: Telos is decentralized in a way that's more like bitcoin than other blockchains (but without the whales who can manipulate price). There was never any single company that started Telos, so there's no company whose CEO could make decisions for the network. There are numerous block producers who decide on any operational issue that isn't clearly described in the TBNOA governance documents. And to get to an action, 15 of the 21 currently active BPs need to sign a multisig transaction. So that's a high threshold. But also, the TBNOA speaks to a large number of issues and so the BPs can't just make up their own rules. Since there are really no whales, no one can vote in any kind of change or bring in their own BPs with their votes. This is also very different from other chains where there are whales. Telos is not located in any one country, so our rules can't be driven by one nation's politics. All in all, this level of decentralization sets Telos apart from almost any blockchain project in existence. People don't have to trust Telos because the system is designed to make trust unnecessary.
Author: Gamals Ahmed, CoinEx Business Ambassador https://preview.redd.it/5bqakdqgl3g51.jpg?width=865&format=pjpg&auto=webp&s=b709794863977eb6554e3919b9e00ca750e3e704 A decentralized storage network that transforms cloud storage into an account market. Miners obtain the integrity of the original protocol by providing data storage and / or retrieval. On the contrary, customers pay miners to store or distribute data and retrieve it. Filecoin announced, that there will be more delays before its main network is officially launched. Filecoin developers postponed the release date of their main network to late July to late August 2020. As mentioned in a recent announcement, the Filecoin team said that the initiative completed the first round of the internal protocol security audit. Platform developers claim that the results of the review showed that they need to make several changes to the protocol’s code base before performing the second stage of the software testing process. Created by Protocol Labs, Filecoin was developed using File System (IPFS), which is a peer-to-peer data storage network. Filecoin will allow users to trade storage space in an open and decentralized market. Filecoin developers implemented one of the largest cryptocurrency sales in 2017. They have privately obtained over $ 200 million from professional or accredited investors, including many institutional investors. The main network was slated to launch last month, but in February 2020, the Philly Queen development team delayed the release of the main network between July 15 and July 17, 2020. They claimed that the outbreak of the Coronavirus (COVID-19) in China was the main cause of the delay. The developers now say that they need more time to solve the problems found during a recent codecase audit. The Filecoin team noted the following: “We have drafted a number of protocol changes to ensure that building our major network launch is safe and economically sound.” The project developers will add them to two different implementations of Filecoin (Lotus and go-filecoin) in the coming weeks. Filecoin developers conducted a survey to allow platform community members to cast their votes on three different launch dates for Testnet Phase 2 and mainnet. The team reported that the community gave their votes. Based on the vote results, the Filecoin team announced a “conservative” estimate that the second phase of the network test should begin by May 11, 2020. The main Filecoin network may be launched sometime between July 20 and August 21, 2020. The updates to the project can be found on the Filecoin Road Map. Filecoin developers stated: “This option will make us get the most important protocol changes first, and then implement the rest as protocol updates during testnet.” Filecoin is back down from the final test stage. Another filecoin decentralized storage network provider launched its catalytic test network, the final stage of the storage network test that supports the blockchain. In a blog post on her website, Filecoin said she will postpone the last test round until August. The company also announced a calibration period from July 20 to August 3 to allow miners to test their mining settings and get an idea of how competition conditions affected their rewards. Filecoin had announced earlier last month that the catalytic testnet test would precede its flagship launch. The delay in the final test also means that the company has returned the main launch window between August 31 and September 21. Despite the lack of clear incentives for miners and multiple delays, Filecoin has succeeded in attracting huge interest, especially in China. Investors remained highly speculating on the network’s mining hardware and its premium price. Mining in Filecoin In most blockchain protocols, “miners” are network participants who do the work necessary to promote and maintain the blockchain. To provide these services, miners are compensated in the original cryptocurrency. Mining in Filecoin works completely differently — instead of contributing to computational power, miners contribute storage capacity to use for dealing with customers looking to store data. Filecoin will contain several types of miners: Storage miners responsible for storing files and data on the network. Miners retrieval, responsible for providing quick tubes for file recovery. Miners repair to be carried out. Storage miners are the heart of the network. They earn Filecoin by storing data for clients, and computerizing cipher directories to check storage over time. The probability of earning the reward reward and transaction fees is proportional to the amount of storage that the Miner contributes to the Filecoin network, not the hash power. Retriever miners are the veins of the network. They earn Filecoin by winning bids and mining fees for a specific file, which is determined by the market value of the said file size. Miners bandwidth and recovery / initial transaction response time will determine its ability to close recovery deals on the network. The maximum bandwidth of the recovery miners will determine the total amount of deals that it can enter into. In the current implementation, the focus is mostly on storage miners, who sell storage capacity for FIL.
The current system specifications recommended for running the miner are:
NVIDIA-manufactured GPU (to be expanded).
SSD drive designated as large buffer (512GB +).
Large amount of RAM for data replication account (128GB +)
Compared to the hardware requirements for running a validity checker, these standards are much higher — although they definitely deserve it. Since these will not increase in the presumed future, the money spent on Filecoin mining hardware will provide users with many years of reliable service, and they pay themselves many times. Think of investing as a small business for cloud storage. To launch a model on the current data hosting model, it will cost millions of dollars in infrastructure and logistics to get started. With Filecoin, you can do the same for a few thousand dollars. Proceed to mining Deals are the primary function of the Filecoin network, and it represents an agreement between a client and miners for a “storage” contract. Once the customer decides to have a miner to store based on the available capacity, duration and price required, he secures sufficient funds in a linked portfolio to cover the total cost of the deal. The deal is then published once the mine accepts the storage agreement. By default, all Filecoin miners are set to automatically accept any deal that meets their criteria, although this can be disabled for miners who prefer to organize their deals manually. After the deal is published, the customer prepares the data for storage and then transfers it to the miner. Upon receiving all the data, the miner fills in the data in a sector, closes it, and begins to provide proofs to the chain. Once the first confirmation is obtained, the customer can make sure the data is stored correctly, and the deal has officially started. Throughout the deal, the miner provides continuous proofs to the chain. Clients gradually pay with money they previously closed. If there is missing or late evidence, the miner is punished. More information about this can be found in the Runtime, Cut and Penalties section of this page. At Filecoin, miners earn two different types of rewards for their efforts: storage fees and reward prevention. Storage fees are the fees that customers pay regularly after reaching a deal, in exchange for storing data. This fee is automatically deposited into the withdrawal portfolio associated with miners while they continue to perform their duties over time, and is locked for a short period upon receipt. Block rewards are large sums given to miners calculated on a new block. Unlike storage fees, these rewards do not come from a linked customer; Instead, the new FIL “prints” the network as an inflationary and incentive measure for miners to develop the chain. All active miners on the network have a chance to get a block bonus, their chance to be directly proportional to the amount of storage space that is currently being contributed to the network. Duration of operation, cutting and penalties “Slashing” is a feature found in most blockchain protocols, and is used to punish miners who fail to provide reliable uptime or act maliciously against the network. In Filecoin, miners are susceptible to two different types of cut: storage error cut, unanimously reduce error. Storage Error Reduction is a term used to include a wider range of penalties, including error fees, sector penalties, and termination fees. Miners must pay these penalties if they fail to provide reliability of the sector or decide to leave the network voluntarily. An error fee is a penalty that a miner incurs for each non-working day. Sector punishment: A penalty incurred by a miner of a disrupted sector for which no error was reported before the WindowPoSt inspection. The sector will pay an error fee after the penalty of the sector once the error is discovered. Termination Fee: A penalty that a miner incurs when a sector is voluntary or involuntarily terminated and removed from the network. Cutting consensus error is the penalty that a miner incurs for committing consensus errors. This punishment applies to miners who have acted maliciously against the network consensus function. Filecoin miners Eight of the top 10 Felticoin miners are Chinese investors or companies, according to the blockchain explorer, while more companies are selling cloud mining contracts and distributed file sharing system hardware. CoinDesk’s Wolfe Chao wrote: “China’s craze for Filecoin may have been largely related to the long-standing popularity of crypto mining in the country overall, which is home to about 65% of the computing power on Bitcoin at discretion.” With Filecoin approaching the launch of the mainnet blocknet — after several delays since the $ 200 million increase in 2017 — Chinese investors are once again speculating strongly about network mining devices and their premium prices. Since Protocol Labs, the company behind Filecoin, released its “Test Incentives” program on June 9 that was scheduled to start in a week’s time, more than a dozen Chinese companies have started selling cloud mining contracts and hardware — despite important details such as economics Mining incentives on the main network are still endless. Sales volumes to date for each of these companies can range from half a million to tens of millions of dollars, according to self-reported data on these platforms that CoinDesk has watched and interviews with several mining hardware manufacturers. Filecoin’s goal is to build a distributed storage network with token rewards to spur storage hosting as a way to drive wider adoption. Protocol Labs launched a test network in December 2019. But the tokens mined in the testing environment so far are not representative of the true silicon coin that can be traded when the main network is turned on. Moreover, the mining incentive economics on testnet do not represent how final block rewards will be available on the main network. However, data from Blockecoin’s blocknetin testnet explorers show that eight out of 10 miners with the most effective mining force on testnet are currently Chinese miners. These eight miners have about 15 petabytes (PB) of effective storage mining power, accounting for more than 85% of the total test of 17.9 petable. For the context, 1 petabyte of hard disk storage = 1000 terabytes (terabytes) = 1 million gigabytes (GB). Filecoin craze in China may be closely related to the long-standing popularity of crypt mining in the country overall, which is home to about 65% of the computing power on Bitcoin by estimation. In addition, there has been a lot of hype in China about foreign exchange mining since 2018, as companies promote all types of devices when the network is still in development. “Encryption mining has always been popular in China,” said Andy Tien, co-founder of 1475, one of several mining hardware manufacturers in Philquin supported by prominent Chinese video indicators such as Fenbushi and Hashkey Capital. “Even though the Velikoyen mining process is more technologically sophisticated, the idea of mining using hard drives instead of specialized machines like Bitcoin ASIC may be a lot easier for retailers to understand,” he said. Meanwhile, according to Feixiaohao, a Chinese service comparable to CoinMarketCap, nearly 50 Chinese crypto exchanges are often somewhat unknown with some of the more well-known exchanges including Gate.io and Biki — have listed trading pairs for Filecoin currency contracts for USDT. In bitcoin mining, at the current difficulty level, one segment per second (TH / s) fragmentation rate is expected to generate around 0.000008 BTC within 24 hours. The higher the number of TH / s, the greater the number of bitcoins it should be able to produce proportionately. But in Filecoin, the efficient mining force of miners depends on the amount of data stamped on the hard drive, not the total size of the hard drive. To close data in the hard drive, the Filecoin miner still needs processing power, i.e. CPU or GPU as well as RAM. More powerful processors with improved software can confine data to the hard drive more quickly, so miners can combine more efficient mining energy faster on a given day. As of this stage, there appears to be no transparent way at the network level for retail investors to see how much of the purchased hard disk drive was purchased which actually represents an effective mining force. The U.S.-based Labs Protocol was behind Filecoin’s initial coin offer for 2017, which raised an astonishing $ 200 million. This was in addition to a $ 50 million increase in private investment supported by notable venture capital projects including Sequoia, Anderson Horowitz and Union Square Ventures. CoinDk’s parent company, CoinDk, has also invested in Protocol Labs. After rounds of delay, Protocol Protocols said in September 2019 that a testnet launch would be available around December 2019 and the main network would be rolled out in the first quarter of 2020. The test started as promised, but the main network has been delayed again and is now expected to launch in August 2020. What is Filecoin mining process? Filecoin mainly consists of three parts: the storage market (the chain), the blockecin Filecoin, and the search market (under the chain). Storage and research market in series and series respectively for security and efficiency. For users, the storage frequency is relatively low, and the security requirements are relatively high, so the storage process is placed on the chain. The retrieval frequency is much higher than the storage frequency when there is a certain amount of data. Given the performance problem in processing data on the chain, the retrieval process under the chain is performed. In order to solve the security issue of payment in the retrieval process, Filecoin adopts the micro-payment strategy. In simple terms, the process is to split the document into several copies, and every time the user gets a portion of the data, the corresponding fee is paid. Types of mines corresponding to Filecoin’s two major markets are miners and warehousers, among whom miners are primarily responsible for storing data and block packages, while miners are primarily responsible for data query. After the stable operation of the major Filecoin network in the future, the mining operator will be introduced, who is the main responsible for data maintenance. In the initial release of Filecoin, the request matching mechanism was not implemented in the storage market and retrieval market, but the takeover mechanism was adopted. The three main parts of Filecoin correspond to three processes, namely the stored procedure, retrieval process, packaging and reward process. The following figure shows the simplified process and the income of the miners: The Filecoin mining process is much more complicated, and the important factor in determining the previous mining profit is efficient storage. Effective storage is a key feature that distinguishes Filecoin from other decentralized storage projects. In Filecoin’s EC consensus, effective storage is similar to interest in PoS, which determines the likelihood that a miner will get the right to fill, that is, the proportion of miners effectively stored in the entire network is proportional to final mining revenue. It is also possible to obtain higher effective storage under the same hardware conditions by improving the mining algorithm. However, the current increase in the number of benefits that can be achieved by improving the algorithm is still unknown. It seeks to promote mining using Filecoin Discover Filecoin announced Filecoin Discover — a step to encourage miners to join the Filecoin network. According to the company, Filecoin Discover is “an ever-growing catalog of numerous petabytes of public data covering literature, science, art, and history.” Miners interested in sharing can choose which data sets they want to store, and receive that data on a drive at a cost. In exchange for storing this verified data, miners will earn additional Filecoin above the regular block rewards for storing data. Includes the current catalog of open source data sets; ENCODE, 1000 Genomes, Project Gutenberg, Berkley Self-driving data, more projects, and datasets are added every day. Ian Darrow, Head of Operations at Filecoin, commented on the announcement: “Over 2.5 quintillion bytes of data are created every day. This data includes 294 billion emails, 500 million tweets and 64 billion messages on social media. But it is also climatology reports, disease tracking maps, connected vehicle coordinates and much more. It is extremely important that we maintain data that will serve as the backbone for future research and discovery”. Miners who choose to participate in Filecoin Discover may receive hard drives pre-loaded with verified data, as well as setup and maintenance instructions, depending on the company. The Filecoin team will also host the Slack (fil-Discover-support) channel where miners can learn more. Filecoin got its fair share of obstacles along the way. Last month Filecoin announced a further delay before its main network was officially launched — after years of raising funds. In late July QEBR (OTC: QEBR) announced that it had ceded ownership of two subsidiaries in order to focus all of the company’s resources on building blockchain-based mining operations. The QEBR technology team previously announced that it has proven its system as a Filecoin node valid with CPU, GPU, bandwidth and storage compatibility that meets all IPFS guidelines. The QEBR test system is connected to the main Filecoin blockchain and the already mined filecoin coin has already been tested. “The disclosure of Sheen Boom and Jihye will allow our team to focus only on the upcoming global launch of Filecoin. QEBR branch, Shenzhen DZD Digital Technology Ltd. (“ DZD “), has a strong background in blockchain development, extraction Data, data acquisition, data processing, data technology research. We strongly believe Filecoin has the potential to be a leading blockchain-based cryptocurrency and will make every effort to make QEBR an important player when Mainecoin mainnet will be launched soon”. IPFS and Filecoin Filecoin and IPFS are complementary protocols for storing and sharing data in a decentralized network. While users are not required to use Filecoin and IPFS together, the two combined are working to resolve major failures in the current web infrastructure. IPFS It is an open source protocol that allows users to store and transmit verifiable data with each other. IPFS users insist on data on the network by installing it on their own device, to a third-party cloud service (known as Pinning Services), or through community-oriented systems where a group of individual IPFS users share resources to ensure the content stays live. The lack of an integrated catalytic mechanism is the challenge Filecoin hopes to solve by allowing users to catalyze long-term distributed storage at competitive prices through the storage contract market, while maintaining the efficiency and flexibility that the IPFS network provides. Using IPFS In IPFS, the data is hosted by the required data installation nodes. For data to persist while the user node is offline, users must either rely on their other peers to install their data voluntarily or use a central install service to store data. Peer-to-peer reliance caching data may be a good thing as one or multiple organizations share common files on an internal network, or where strong social contracts can be used to ensure continued hosting and preservation of content in the long run. Most users in an IPFS network use an installation service. Using Filecoin The last option is to install your data in a decentralized storage market, such as Filecoin. In Filecoin’s structure, customers make regular small payments to store data when a certain availability, while miners earn those payments by constantly checking the integrity of this data, storing it, and ensuring its quick recovery. This allows users to motivate Filecoin miners to ensure that their content will be live when it is needed, a distinct advantage of relying only on other network users as required using IPFS alone. Filecoin, powered by IPFS It is important to know that Filecoin is built on top of IPFS. Filecoin aims to be a very integrated and seamless storage market that takes advantage of the basic functions provided by IPFS, they are connected to each other, but can be implemented completely independently of each other. Users do not need to interact with Filecoin in order to use IPFS. Some advantages of sharing Filecoin with IPFS:
Filecoin and IPFS CIDs share hash specifications.
Use libp2p by Filecoin nodes to create secure connections with each other.
Messaging between nodes and cluster propagation is facilitated in Filecoin by libp2p pubsub.
IPLD use for blockchain data structures.
Use Graphsync to transfer data between nodes.
Of all the decentralized storage projects, Filecoin is undoubtedly the most interested, and IPFS has been running stably for two years, fully demonstrating the strength of its core protocol. Filecoin’s ability to obtain market share from traditional central storage depends on end-user experience and storage price. Currently, most Filecoin nodes are posted in the IDC room. Actual deployment and operation costs are not reduced compared to traditional central cloud storage, and the storage process is more complicated. PoRep and PoSt, which has a large number of proofs of unknown operation, are required to cause the actual storage cost to be so, in the early days of the release of Filecoin. The actual cost of storing data may be higher than the cost of central cloud storage, but the initial storage node may reduce the storage price in order to obtain block rewards, which may result in the actual storage price lower than traditional central cloud storage. In the long term, Filecoin still needs to take full advantage of its P2P storage, convert storage devices from specialization to civil use, and improve its algorithms to reduce storage costs without affecting user experience. The storage problem is an important problem to be solved in the blockchain field, so a large number of storage projects were presented at the 19th Web3 Summit. IPFS is an important part of Web3 visibility. Its development will affect the development of Web3 to some extent. Likewise, Web3 development somewhat determines the future of IPFS. Filecoin is an IPFS-based storage class project initiated by IPFS. There is no doubt that he is highly expected. Resources :
Diamond in the rough forks to Blockcore tech. Solid 10x short term.
Meet x42 Protocol, the feeless blockchain that will allow anyone to host decentralized applications the easy way. X42 Protocol Price $0.01275764 Market Cap $247,173 (May 27, 2020) Explorer: https://explorer.x42.tech/ Website: https://www.x42.tech/ x42 is a decentralized cryptocurrency based on the Blockcore technology designed to be a multi-chain solution for DApps (decentralized applications) that allows for a range that goes anywhere from small indie developed games to large entrepreneurial projects that span dozens of facilities. The main idea behind x42 is to be a scalable, on-chain solution for any developer that wants to launch games and applications in general with minimal initial investment, zero transaction fees and near infinite scalability. The protocol works around a main blockchain which hosts all the x42 coins and three types of nodes. Side blockchains can be created at will by the developers that decide to use the x42 protocol to launch their projects, side blockchains are very flexible and allow for a great deal of customization. The main blockchain of the x42 protocol will have a maximum total of forty-two million coins mined into existence by the year 2030. The coins follow the same rules as most cryptocurrencies, every transaction is final, timestamped and will be registered to the blockchain ledger, blockchain explorers can be used to browse any and all transactions Every project can have its own side blockchain, in which the development team can fully test and experiment on before going live to the store, because of that need all side blockchains have access to individual testnets. FM Interview with developer: The workpaper talks about using Stratis technology - Many say this is soon to be defunct? It does, the current WP is the 1st itteration, we have a 2.0 that will be released when the servers go online. However in Dec 2019, we have switched to Blockcore as Stratis had too many limitations that didn't meet our needs without heavy modifications. We did start with stratis as the basic framework, but we modified all of the code thus far to meet the needs we have. x42 is not stratis, but did start with some of their code. The switch to blockcore is really more of a code thing and not otherwise noticed on the user end. There has been some but not a significant amount of discussion about it. Anyone that does access the new beta will be able to notice the difference as it nolonger will say stratis anywhere. On the user end, some file structures may change with mention of blockcore. Blockcore is a fork of Stratis, however the cool thing is that it isnt its own project. More details: https://www.blockcore.net/ is an open source project started by a group of blockchain developers and engineers to realize a fully integrated platform for building custom blockchains. Blockcore is the foundation for realizing blockchains and includes core functionality to create your own custom blockchain with a lot of tooling supporting your blockchain. So its been 18 months of development? When will we say a mainnet? We are currently in testnet and close to 90% ready. Xservers should be released this year. However we are not working on a timeline for mainnet with dates at the moment. Deadlines produce inferior products. The new wallet (xCore) has passed and is ready, it will work alongside the xServer. The switch to blockcore did require some code changes and we are still working with blockcore to fine tune some issues in their code as well. Security and ease of use for the end user are required. What about use of dapps in beta? We have a few in progress, but there are not many as the parameters can be changing. Although, anyone that can code a dapp in c# could drop it into their app folder and have one right now. There just won't be the user interface on the current wallet or the server support yet. Why did you call it x42? X has always been the universal unknown in mathematics, which can also mean ‘anything.’ X is believed to have come from the arab letter shin (ش), that was initially used by spanish scholars because of their inability to translate certain arabic sounds into the spanish language, so it became a synonym for an unknown thing. From there on it was noticed that the spanish language didn’t have an appropriate sound for the arabic ‘sh.’ As time passed it was adapted into the ‘ck’ sound, which in classic greek is written down with a symbol known as chi (X). 42 is known to Douglas Adams’ The Hitchhiker’s Guide to the Galaxy readers as the ‘answer to the ultimate question of life, the universe, and everything’. We believe that x42 can offer a great answer to most questions plaguing the cryptocurrency space at the moment and in the future. That mix of an unknown and an answer is where our name got its origin. The techy Details: The two main levels of the x42 architecture are: xCore – This is responsible for interacting with the infrastructure and the interface layers in addition to the node policy layers. An x42 xCore full node handles all the APIs and user interfaces. NBitcoin – This is responsible for handling messages between nodes in the Network Layer. It also bridges the gap between various Consensus Layer functions. Higher Transaction Speed: Transaction on the x42 blockchain platform happens almost instantly under 60 seconds. Also, the main x42 blockchain supports around 70 transactions per second. This is nearly 10 times the transactions supported by the Bitcoin blockchain network. Additionally, the good thing is the x42 side blockchains can be edited to hold as many transactions-per-second (TPS) as the developers want. Understanding xCore, xServer, and Wallets xCore – An xCore node is basically a device with software that has the entire x42 blockchain saved and connects the main blockchain to the side blockchains. This node seeds the main and side blockchain to all users and also has all the functionalities of a client node. Client node is any client running a wallet connected to the x42 main blockchain network. As explained in the white-paper, “The xCore can stake coins on the main blockchain, it can also run decentralised applications hosted on xServers all across the network after having it installed on the local machine”. xServer – The x42 blockchain supports several different types of servers. xServer is a special sort of wallet that runs on dedicated hardware facilitates decentralisation of applications (DApps), processing and data storage. xServers are also tasked with other activities on the network like propagating smart contracts and signing transactions. It also allows the server owner (the Gatekeeper) to get paid for hosting diverse applications. Besides, xServers can choose to host the projects launched on any of the non-private side blockchains. Projects that share a considerable part of their revenues with the xServer owners are likely to get hosted quickly and maintained by the same servers for a longer period of time. Before we understand xServer wallets, let’s brief about the Hot Wallet and Cold Wallets. A hot wallet is an online wallet always connected to the internet. This is an instant point-of-contact for receiving and sending payments. The hot wallets receive the staking rewards along with payouts from DApps, side blockchain, smart contracts, and private transactions. A cold wallet is an offline wallet and more secure from the hot wallet as it is less susceptible to online attacks. This wallet holds the collateral for the server. All the server rewards are paid out to the cold storage wallet. Coming to the xServer wallets, these are basically software holding information for both hot and cold storage wallets. The xServer wallets can run on almost any operating system. They can also host and launch decentralised applications (DApps) and generate passive income by DApp hosting. Nice article.. https://medium.com/the-consensus/x42-protocol-not-just-another-pos-mn-project-d8dd73e8846d
Dieter Fishbein, Ecosystem Development Lead, Web3 Foundation
Dan: Hey everyone, thanks for joining us for the Polkadot Launch AMA. We have Dieter Fishbein (Head of Ecosystem Development, our business development team), Logan Saether (Technical Education), and Will Pankiewicz (Master of Validators) joining us today. We had some great questions submitted in advance, and we’ll start by answering those and learning a bit about each of our guests. After we go through the pre-submitted questions, then we’ll open up the chat to live Q&A and the hosts will answer as many questions as they can. We’ll start off with Dieter and ask him a set of some business-related questions.
Dieter could you introduce yourself, your background, and your role within the Polkadot ecosystem?
Dieter: I got my start in the space as a cryptography researcher at the University of Waterloo. This is where I first learned about Bitcoin and started following the space. I spent the next four years or so on the investment team for a large asset manager where I primarily focused on emerging markets. In 2017 I decided to take the plunge and join the space full-time. I worked at a small blockchain-focused VC fund and then joined the Polkadot team just over a year ago. My role at Polkadot is mainly focused on ensuring there is a vibrant community of projects building on our technology.
Q: Adoption of Polkadot of the important factors that all projects need to focus on to become more attractive to the industry. So, what is Polkadot's plan to gain more Adoption? [sic]
A (Dieter): Polkadot is fundamentally a developer-focused product so much of our adoption strategy is focused around making Polkadot an attractive product for developers. This has many elements. Right now the path for most developers to build on Polkadot is by creating a blockchain using the Substrate framework which they will later connect to Polkadot when parachains are enabled. This means that much of our adoption strategy comes down to making Substrate an attractive tool and framework. However, it’s not just enough to make building on Substrate attractive, we must also provide an incentive to these developers to actually connect their Substrate-based chain to Polkadot. Part of this incentive is the security that the Polkadot relay chain provides but another key incentive is becoming interoperable with a rich ecosystem of other projects that connect to Polkadot. This means that a key part of our adoption strategy is outreach focused. We go out there and try to convince the best projects in the space that building on our technology will provide them with significant value-add. This is not a purely technical argument. We provide significant support to projects building in our ecosystem through grants, technical support, incubatoaccelerator programs and other structured support programs such as the Substrate Builders Program (https://www.substrate.io/builders-program). I do think we really stand out in the significant, continued support that we provide to builders in our ecosystem. You can also take a look at the over 100 Grants that we’ve given from the Web3 Foundation: https://medium.com/web3foundation/web3-foundation-grants-program-reaches-100-projects-milestone-8fd2a775fd6b
Q: On moving forward through your roadmap, what are your most important next priorities? Does the Polkadot team have enough fundamentals (Funds, Community, etc.) to achieve those milestones?
A (Dieter): I would say the top priority by far is to ensure a smooth roll-out of key Polkadot features such as parachains, XCMP and other key parts of the protocol. Our recent Proof of Authority network launch was only just the beginning, it’s crucial that we carefully and successfully deploy features that allow builders to build meaningful technology. Second to that, we want to promote adoption by making more teams aware of Polkadot and how they can leverage it to build their product. Part of this comes down to the outreach that I discussed before but a major part of it is much more community-driven and many members of the team focus on this. We are also blessed to have an awesome community to make this process easier 🙂
Q: Where can a list of Polkadot's application-specific chains can be found?
A (Dieter): The best list right now is http://www.polkaproject.com/. This is a community-led effort and the team behind it has done a terrific job. We’re also working on providing our own resource for this and we’ll share that with the community when it’s ready.
Q: Could you explain the differences and similarities between Kusama and Polkadot?
A (Dieter): Kusama is fundamentally a less robust, faster-moving version of Polkadot with less economic backing by validators. It is less robust since we will be deploying new technology to Kusama before Polkadot so it may break more frequently. It has less economic backing than Polkadot, so a network takeover is easier on Kusama than on Polkadot, lending itself more to use cases without the need for bank-like security. In exchange for lower security and robustness, we expect the cost of a parachain lease to be lower on Kusama than Polkadot. Polkadot will always be 100% focused on security and robustness and I expect that applications that deal with high-value transactions such as those in the DeFi space will always want a Polkadot deployment, I think there will be a market for applications that are willing to trade cheap, high throughput for lower security and robustness such as those in the gaming, content distribution or social networking sectors. Check out - https://polkadot.network/kusama-polkadot-comparing-the-cousins/ for more detailed info!
Q: and for what reasons would a developer choose one over the other?
A (Dieter): Firstly, I see some earlier stage teams who are still iterating on their technology choosing to deploy to Kusama exclusively because of its lower-stakes, faster moving environment where it will be easier for them to iterate on their technology and build their user base. These will likely encompass the above sectors I identified earlier. To these teams, Polkadot becomes an eventual upgrade path for them if, and when, they are able to perfect their product, build a larger community of users and start to need the increased stability and security that Polkadot will provide. Secondly, I suspect many teams who have their main deployment on Polkadot will also have an additional deployment on Kusama to allow them to test new features, either their tech or changes to the network, before these are deployed to Polkadot mainnet.
Logan Saether, Technical Education, Web3 Foundation
Q: Sweet, let's move over to Logan. Logan - could you introduce yourself, your background, and your role within the Polkadot ecosystem?
A (Logan): My initial involvement in the industry was as a smart contract engineer. During this time I worked on a few projects, including a reboot of the Ethereum Alarm Clock project originally by Piper Merriam. However, I had some frustrations at the time with the limitations of the EVM environment and began to look at other tools which could help me build the projects that I envisioned. This led to me looking at Substrate and completing a bounty for Web3 Foundation, after which I applied and joined the Technical Education team. My responsibilities at the Technical Education team include maintaining the Polkadot Wiki as a source of truth on the Polkadot ecosystem, creating example applications, writing technical documentation, giving talks and workshops, as well as helping initiatives such as the Thousand Validator Programme.
Q: The first technical question submitted for you was: "When will an official Polkadot mobile wallet appear?"
A (Logan): There is already an “official” wallet from Parity Technologies called the Parity Signer. Parity Signer allows you to keep your private keys on an air-gapped mobile device and to interactively sign messages using web interfaces such as Polkadot JS Apps. If you’re looking for something that is more of an interface to the blockchain as well as a wallet, you might be interested in PolkaWallet which is a community team that is building a full mobile interface for Polkadot. For more information on Parity Signer check out the website: https://www.parity.io/signe
Q: Great thanks...our next question is: If someone already developed an application to run on Ethereum, but wants the interoperability that Polkadot will offer, are there any advantages to rebuilding with Substrate to run as a parachain on the Polkadot network instead of just keeping it on Ethereum and using the Ethereum bridge for use with Polkadot?
A (Logan): Yes, the advantage you would get from building on Substrate is more control over how your application will interact with the greater Polkadot ecosystem, as well as a larger design canvas for future iterations of your application. Using an Ethereum bridge will probably have more cross chain latency than using a Polkadot parachain directly. The reason for this is due to the nature of Ethereum’s separate consensus protocol from Polkadot. For parachains, messages can be sent to be included in the next block with guarantees that they will be delivered. On bridged chains, your application will need to go through more routes in order to execute on the desired destination. It must first route from your application on Ethereum to the Ethereum bridge parachain, and afterward dispatch the XCMP message from the Polkadot side of the parachain. In other words, an application on Ethereum would first need to cross the bridge then send a message, while an application as a parachain would only need to send the message without needing to route across an external bridge.
Q: DOT transfers won't go live until Web3 removes the Sudo module and token holders approve the proposal to unlock them. But when will staking rewards start to be distributed? Will it have to after token transfers unlock? Or will accounts be able to accumulate rewards (still locked) once the network transitions to NPoS?
A (Logan): Staking rewards will be distributed starting with the transition to NPoS. Transfers will still be locked during the beginning of this phase, but reward payments are technically different from the normal transfer mechanism. You can read more about the launch process and steps at http://polkadot.network/launch-roadmap
Q: Next question is: I'm interested in how Cumulus/parachain development is going. ETA for when we will see the first parachain registered working on Kusama or some other public testnet like Westend maybe?
A (Logan): Parachains and Cumulus is a current high priority development objective of the Parity team. There have already been PoC parachains running with Cumulus on local testnets for months. The current work now is making the availability and validity subprotocols production ready in the Polkadot client. The best way to stay up to date would be to follow the project boards on GitHub that have delineated all of the tasks that should be done. Ideally, we can start seeing parachains on Westend soon with the first real parachains being deployed on Kusama thereafter. The projects board can be viewed here: https://github.com/paritytech/polkadot/projects Dan: Also...check out Basti's tweet from yesterday on the Cumulus topic: https://twitter.com/bkchstatus/1270479898696695808?s=20
Q: In what ways does Polkadot support smart contracts?
A (Logan): The philosophy behind the Polkadot Relay Chain is to be as minimal as possible, but allow arbitrary logic at the edges in the parachains. For this reason, Polkadot does not support smart contracts natively on the Relay Chain. However, it will support smart contracts on parachains. There are already a couple major initiatives out there. One initiative is to allow EVM contracts to be deployed on parachains, this includes the Substrate EVM module, Parity’s Frontier, and projects such as Moonbeam. Another initiative is to create a completely new smart contract stack that is native to Substrate. This includes the Substrate Contracts pallet, and the ink! DSL for writing smart contracts. Learn more about Substrate's compatibility layer with Ethereum smart contracts here: https://github.com/paritytech/frontier
Will Pankiewicz, Master of Validators, Parity Technologies
Q: (Dan) Thanks for all the answers. Now we’ll start going through some staking questions with Will related to validating and nominating on Polkadot. Will - could you introduce yourself, your background, and your role within the Polkadot ecosystem?
A (Will): Sure thing. Like many others, Bitcoin drew me in back in 2013, but it wasn't until Ethereum came that I took the deep dive into working in the space full time. It was the financial infrastructure aspects of cryptocurrencies I was initially interested in, and first worked on dexes, algorithmic trading, and crypto funds. I really liked the idea of "Generalized Mining" that CoinFund came up with, and started to explore the whacky ways the crypto funds and others can both support ecosystems and be self-sustaining at the same time. This drew me to a lot of interesting experiments in what later became DeFi, as well as running validators on Proof of Stake networks. My role in the Polkadot ecosystem as “Master of Validators” is ensuring the needs of our validator community get met.
Q: Cool thanks. Our first community question was "Is it still more profitable to nominate the validators with lesser stake?"
A (Will): It depends on their commission, but generally yes it is more profitable to nominate validators with lesser stake. When validators have lesser stake, when you nominate them this makes your nomination stake a higher percentage of total stake. This means when rewards get distributed, it will be split more favorably toward you, as rewards are split by total stake percentage. Our entire rewards scheme is that every era (6 hours in Kusama, 24 hours in Polkadot), a certain amount of rewards get distributed, where that amount of rewards is dependent on the total amount of tokens staked for the entire network (50% of all tokens staked is currently optimal). These rewards from the end of an era get distributed roughly equally to all validators active in the validator set. The reward given to each validator is then split between the validators and all their nominators, determined by the total stake that each entity contributes. So if you contribute to a higher percentage of the total stake, you will earn more rewards.
Q: What does priority ranking under nominator addresses mean? For example, what does it mean that nominator A has priority 1 and nominator B has priority 6?
A (Will): Priority ranking is just the index of the nomination that gets stored on chain. It has no effect on how stake gets distributed in Phragmen or how rewards get calculated. This is only the order that the nominator chose their validators. The way that stake from a nominator gets distributed from a nominator to validators is via Phragmen, which is an algorithm that will optimally put stake behind validators so that distribution is roughly equal to those that will get in the validator set. It will try to maximize the total amount at stake in the network and maximize the stake behind minimally staked validators.
Q: On Polkadot.js, what does it mean when there are nodes waiting on Polkadot?
**A (Will):**In Polkadot there is a fixed validator set size that is determined by governance. The way validators get in the active set is by having the highest amount of total stake relative to other validators. So if the validator set size is 100, the top 100 validators by total stake will be in the validator set. Those not active in the validator set will be considered “waiting”.
Q: Another question...Is it necessary to become a waiting validator node right now?
A (Will): It's not necessary, but highly encouraged if you actively want to validate on Polkadot. The longer you are in the waiting tab, the longer you get exposure to nominators that may nominate you.
Q: Will current validators for Kusama also validate for Polkadot? How strongly should I consider their history (with Kusama) when looking to nominate a good validator for DOTs?
A (Will): A lot of Kusama validators will also be validators for Polkadot, as KSM was initially distributed to DOT holders. The early Kusama Validators will also likely be the first Polkadot validators. Being a Kusama validator should be a strong indicator for who to nominate on Polkadot, as the chaos that has ensued with Kusama has allowed validators to battle test their infrastructure. Kusama validators by now are very familiar with tooling, block explorers, terminology, common errors, log formats, upgrades, backups, and other aspects of node operation. This gives them an edge against Polkadot validators that may be new to the ecosystem. You should strongly consider well known Kusama validators when making your choices as a nominator on Polkadot.
Q: Can you go into more details about the process for becoming a DOT validator? Is it similar as the KSM 1000 validators program?
A (Will): The Process for becoming a DOT validators is first to have DOTs. You cannot be a validator without DOTs, as DOTs are used to pay transaction fees, and the minimum amount of DOTs you need is enough to create a validate transaction. After obtaining enough DOTs, you will need to set up your validator infrastructure. Ideally you should have a validator node with specs that match what we call standard hardware, as well as one or more sentry nodes to help isolate the validator node from attacks. After the infrastructure is up and running, you should have your Polkadot accounts set up right with a stash bonded to a controller account, and then submit a validate transaction, which will tell the network your nodes are ready to be a part of the network. You should then try and build a community around your validator to let others know you are trustworthy so that they will nominate you. The 1000 validators programme for Kusama is a programme that gives a certain amount of nominations from the Web3 Foundation and Parity to help bootstrap a community and reputation for validators. There may eventually be a similar type of programme for Polkadot as well. Dan: Thanks a lot for all the answers, Will. That’s the end of the pre-submitted questions and now we’ll open the chat up to live Q&A, and our three team members will get through as many of your questions as possible. We will take questions related to business development, technology, validating, and staking. For those wondering about DOT: DOT tokens do not exist yet. Allocations of Polkadot's native DOT token are technically and legally non-transferable. Hence any publicized sale of DOTs is unsanctioned by Web3 Foundation and possibly fraudulent. Any official public sale of DOTs will be announced on the Web3 Foundation website. Polkadot’s launch process started in May and full network decentralization later this year, holders of DOT allocations will determine issuance and transferability. For those who participated in previous DOT sales, you can learn how to claim your DOTs here (https://wiki.polkadot.network/docs/en/claims).
Telegram Community Follow-up Questions Addressed Below
Q: Polkadot looks good but it confuses me that there are so many other Blockchain projects. What should I pay attention in Polkadot to give it the importance it deserves? What are your planning to achieve with your project?
A (Will): Personally, what I think differentiates it is the governance process. Coordinating forkless upgrades and social coordination helps stand it apart. A (Dieter): The wiki is awesome - https://wiki.polkadot.network/
Q: Over 10,000 ETH paid as a transaction fee , what if this happens on Polkadot? Is it possible we can go through governance to return it to the owner?
Q: What is the minimum ideal amount of DOT and KSM to have if you want to become a validator and how much technical knowledge do you need aside from following the docs?
A (Will): It depends on what the other validators in the ecosystem are staking as well as the validator set size. You just need to be in the top staking amount of the validator set size. So if its 100 validators, you need to be in the top 100 validators by stake.
Q: Will Web3 nominate validators? If yes, which criteria to be elected?
Q: How did you manage to prove that the consensus protocol is safe and unbreakable mathematically?
A (Dieter): We have a research teams of over a dozen scientists with PhDs and post-docs in cryptography and distributed computing who do thorough theoretical analyses on all the protocols used in Polkadot
Q: What are the prospects for NFT?
A: Already being built 🙂
Q: What will be Polkadot next roadmap for 2020 ?
A (Dieter): Building. But seriously - we will continue to add many more features and upgrades to Polkadot as well as continue to strongly focus on adoption from other builders in the ecosystem 🙂 A (Will): https://polkadot.network/launch-roadmap/ This is the launch roadmap. Ideally adding parachains and xcmp towards the end of the year
Q: How Do you stay active in terms of marketing developments during this PANDEMIC? Because I'm sure you're very excited to promote more after this settles down.
A (Dan): The main impact of covid was the impact on in-person events. We have been very active on Crowdcast for webinars since 2019, so it was quite the smooth transition to all-online events. You can see our 40+ past event recordings and follow us on Crowdcast here: https://www.crowdcast.io/polkadot. If you're interested in following our emails for updates (including online events), subscribe here: https://info.polkadot.network/subscribe
Q: Hi, who do you think is your biggest competitor in the space?
A (Dan): Polkadot is a metaprotocol that hasn't been seen in the industry up until this point. We hope to elevate the industry by providing interoperability between all major public networks as well as private blockchains.
Q: Is Polkadot a friend or competitor of Ethereum?
A: Polkadot aims to elevate the whole blockchain space with serious advancements in interoperability, governance and beyond :)
Q: When will there be hardware wallet support?
A (Will): Parity Signer works well for now. Other hardware wallets will be added pretty soon
Q: What are the attractive feature of DOT project that can attract any new users ?
A: We are working on integrations on all the biggest and best wallet providers. ;)
Q: What event/call can we track to catch a switch to nPOS? Is it only force_new_era call? Thanks.
A (Will): If you're on riot, useful channels to follow for updates like this are #polkabot:matrix.org and #polkadot-announcements:matrix.parity.io A (Logan): Yes this is the trigger for initiating the switch to NPoS. You can also poll the ForceEra storage for when it changes to ForceNew.
Q: What strategy will the Polkadot Team use to make new users trust its platform and be part of it?
Q: What problems do you see occurring in the blockchain industry nowadays and how does your project aims to solve these problems?
A (Will): Governance I see as a huge problem. For example upgrading Bitcoin and making decisions for changing things is a very challenging process. We have robust systems of on-chain governance to help solve these coordination problems
Q: How involved are the Polkadot partners? Are they helping with the development?
Q: Can you explain the role of the treasury in Polkadot?
A (Will): The treasury is for projects or people that want to build things, but don't want to go through the formal legal process of raising funds from VCs or grants or what have you. You can get paid by the community to build projects for the community. A: There’s a whole section on the wiki about the treasury and how it functions here https://wiki.polkadot.network/docs/en/mirror-learn-treasury#docsNav
Q: Any plan to introduce Polkadot on Asia, or rising market on Asia?
**A (Will):**We're globally focused
Q: What kind of impact do you expect from the Council? Although it would be elected by token holders, what kind of people you wish to see there?
A (Will): Community focused individuals like u/jam10o that want to see cool things get built and cool communities form If you have further questions, please ask in the official Polkadot Telegram channel.
2020-05-19T20:51:34Z New outbound peer connected: version: 70015, blocks=1378460, peer=0 2020-05-19T20:51:35Z New outbound peer connected: version: 70015, blocks=1364809, peer=1 2020-05-19T20:51:35Z New outbound peer connected: version: 70015, blocks=1379769, peer=2 2020-05-19T20:51:47Z New outbound peer connected: version: 70015, blocks=1378460, peer=4 2020-05-19T20:51:53Z New outbound peer connected: version: 70015, blocks=1327928, peer=6 2020-05-19T20:52:05Z New outbound peer connected: version: 70015, blocks=1364809, peer=7 2020-05-19T20:52:06Z New outbound peer connected: version: 70015, blocks=1327928, peer=8 2020-05-19T20:52:08Z New outbound peer connected: version: 70015, blocks=1379769, peer=9
Seeing some problems with block 1,348,461 being marked invalid:
2020-05-19T21:08:41Z ERROR: invalid header received 2020-05-19T21:08:41Z ERROR: AcceptBlockHeader: block 0000000099f5509b5f36b1926bcf82b21d936ebeadee811030dfbbb7fae915d7 is marked invalid
Block in question: http://testnet.imaginary.cash/block/0000000099f5509b5f36b1926bcf82b21d936ebeadee811030dfbbb7fae915d7 Using Bitcoin ABC Daemon version v0.21.7-unk on 2 separate testnet nodes, hitting the same invalid block EDIT: Don't know what happened with testnet as the BitcoinABC daemon updates itself nightly, but a block was marked invalid on both nodes. The fix for me: bitcoin-cli -rpcconnect=10.0.0.1 -rpcport=8000 --testnet --rpccookiefile=/my/cookie/is/here/.cookie reconsiderblock 0000000099f5509b5f36b1926bcf82b21d936ebeadee811030dfbbb7fae915d7
If you have decided to read all this, thanks, keep reading for a concise breakdown!
So what's the current big thing going on with ARK right NOW?
ARK.io has recently announced on both its blog and its Twitter that ARK Core v2.6 is coming to Mainnet February 11th. The iteration of 2.6 may sound anticlimactic, but it's far from that. Core v2.6 is the biggest upgrade to date- even bigger than the total Core overhaul performed for v2.0, deployed late 2018. The new version brings new transaction types to the ARK Public Network, including types that will play a role in creating an ecosystem of linked chains. This ecosystem of linked chains will have the ARK Public Network in the center of the action, storing chain details and allowing for chain discovery. These new transaction types include: Multipayments — sending to multiple ARK addresses, while just initiating one transaction, saves time and cost Multisignatures — you can now get all of the benefits of multisignatures where more than one user can propose or spend funds depending on the predefined terms (eg. 2 out of 3 users needed to successfully send tokens, vote, …) IPFS — register IPFS compliant hashes on the ARK blockchain within Desktop Wallet. Business & Bridgechain registrations — you can now register your business and bridgechain on the blockchain and soon, you will be able to get verified via our Marketplace to get access to some exciting new features. Delegate resignation — delegates who don’t want to be voted for anymore can now opt-out of this by simply initiating delegate resignation. Additionally, the Core v2.6 improves security against double-spend attacks by implementing nonces. Also, massive enhancements were made to the GTI or Generic Transaction Interface, a critical tool for developers who wish to develop decentralized applications.
What is ARK's unique approach to current issues plaguing the blockchain industry?
ARK empowers everyone, regardless of their aim or technical background, to quickly and easily leverage blockchain technology. In the current hype-driven blockchain landscape, ARK acts as a beacon for individuals, groups, and enterprises seeking to apply blockchain technology to both reach their individual goals and affect change in their local community. ARK’s uniquely simple technology stack allows almost anyone to create and deploy standalone blockchains for any use case with an intuitive graphical user interface and experience. These newly created blockchains also known as bridgechains will have the ability to interoperate through ARK SmartBridge Technology. ARK is also reinventing smart contracts with ARK Logic, a collection of tools including custom transaction types, templates, and plugins. ARK Logic brings security, adaptability, and scalability to decentralized computing workflows. Most importantly, the ARK Ecosystem fosters a growing international community of developers, node operators, blockchains, businesses, and enthusiasts who collectively breathe life into this disruptive technology. Get into the interactive whitepaper here.
Tell me about the ARK Public Network
Ok, no problem. Since coming online on March 21, 2017, the APN has operated as a P2P cryptocurrency with fast block times of 8 seconds and low dynamic fees (near a penny and somewhat novel for a DPoS blockchain). However, the end goal of the APN far exceeds that of just a cryptocurrency that is faster and cheaper to use than Bitcoin. I'll explain further in a minute. The network, as mentioned, is set up as Delegated Proof-of-Stake. This means forging delegates are deemed worthy to secure the chain and add blocks to it by the holders of the ARK token, which vote for delegates using their ARK as vote weight. ARK remains in users' control at all times, and the top 51 delegates in vote weight enter forging status. The network awards each delegate 2 ARK per block (~12,705 ARK/mo) for services rendered. This continues ad infinitum resulting in a declining inflation rate each year (relative to total supply). When users add or remove ARK from a voting wallet address, vote weight adjusts automatically and they don't need to vote again. Voting continues even if user's wallet is offline. The main uses of ARK as the cryptoasset of the ARK Public Network besides being a P2P cryptocurrency include:
Being a medium of exchange for ARK Public Network services. Delegates and businesses can operate services where transactions are settled in ARK.
Allowing convenience in interoperability, giving users access to bridgechain use cases via the main ARK cryptoasset.
Liquidity for small and large ARK bridgechains. This is via ARK SmartBridge Technology/ARK Swap Market (in development)
Payment method for plugin, delegate, and talent marketplaces. Services rendered can have transactions settled using ARK.
Yes, team good. Team very good. General sentiment among ARK team members is that ARK is a dream project to work on, and this motivates them to do great work on a consistent basis as the ARK technology stack progresses. Very recently, ARK hired an additional half dozen people in various departments, including marketing department. This brings ARK team total to over three dozen experts. The ARK business entity is also well funded with around 10 years of budget. The ARK business entity spends funds in a very sensible manner compared to some other projects who spend with insufficient foresight or discretion. Members of the board are thoughtful and deliberate, and the CEO FX Thoorens has been hard at work putting a spotlight on ARK, showing an 'intermeshing' of ARK with the global regulatory landscape in regards to crypto. Recently, ARK became a founding member of ADAN, a professional association based in France created to help structure and develop the digital assets industry. Other members include Consensys France and Ledger. ADAN will consult with public authorities, industry leaders and private bodies to promote the use of digital assets and all activities in this sector. This includes exchange platforms, brokers, hardware, protocols, decentralized applications and blockchain technology platforms. Hear FX Thoorens talk more about this in this podcast episode. The ARK business entity is located in France, but the ARK team is distributed across 10+ countries and multiple continents.
What's going to happen?
Cool stuff. Organizations and open source projects have been stumbling across ARK and really like what they see. Multiple projects are working with ARK technology and are at various stages of development, but since you're busy, I'll highlight the project nOS which recently launched their public testnet and uses ARK technology for their blockchain. nOS also has great things to say about ARK that you can hear in this podcast episode or watch in this video. We believe that as more businesses, organizations, and open source projects start looking around for blockchain solutions, they will also enjoy ARK's simplicity, flexibility, and feature set. Our powerful technology stack is backed up by a recently upgraded documentation hub for developers. The product we have that makes it very easy for projects to join the ARK Ecosystem is called the ARK Deployer, which you can learn about in this two minute video. It allows developers from all walks of life to create, customize and launch a standalone blockchain in three simple steps. In the near future, what's going to happen is a big improvement to the Deployer. The ARK Deployer will get an upgraded and more powerful user interface that also facilitates chain management post-launch, as well as interface directly with cloud providers like Digital Ocean to launch genesis node and peers in background. This would allow for a massive leap forward in our vision of 'Point. Click. Blockchain.' ARK.io is also working on a Marketplace for developers, where custom plugins and tools developed by both ARK.io as well as third parties can be acquired for assembling blockchains much easier. Imagine a wordpress-type environment where you can create a super-powerful and customized blockchain by connecting Legos together. In the same way that early World Wide Web needed WordPress/Squarespace style tools to bring the technology to every business or organization, we believe that this need will be out there for blockchain technology as this new decade progresses. There is more cool stuff that is going to happen, but I'll wrap it up there for now.
After reading all this stuff, what is it you want me to do?
Well, not make any financial decisions, because that is not the purpose of this information. However, as a developer, there's a lot of interesting things you should know and may want to consider doing. The ARK technology stack uses TypeScript and other JS-style frameworks, so if you know those, you should get excited.
Earn a lot more ARK. The Tier 0 Program offers bigger projects we need help with and therefore more ARK. You can even contact the team with an idea for a Tier 0 Project you want to do that makes ARK look cool. For example, there is a Tier 0 project designed to highlight ARK tech as a proof of concept for scooter rentals. See program status here.
Look into the ARK Deployer for making your own chain with a custom use case. If you are a part of a project that is currently just a token on someone else's mainnet, and you have scaling concerns or issues with sovereignty, ARK should be a candidate for upgrading your solution for this new decade. Check out ARK Deployer here.
Here's some additional less 'developery' stuff you can do:
Get your idea funded through theARK Community Fund. It's community run, and operated by community elected board members. Your idea can be anything that helps ARK, maybe some seed money for a business like this one that ships ARK Stickers worldwide, or maybe some small dev project, or video production, or article, etc.
Subscribe to the ARK Crypto Podcast. It happens weekly, and it's one of the absolute best podcasts in the space that's centered around a specific project. It knows you're busy, so it's to the point, well constructed, and entertaining. The podcast is looking for subscribers. Do it on iTunes, Google Play, Spotify, Castbox, etc. Here's an updated ARK overview episode that covers much more than you read in this post today.
Other subscribing goodies. Twitter, Slack, Discord, Reddit, Facebook, Medium, etc list of links here.
Just in case, throwing the website URL here, which is ARK.io - it's a really good website that has more information for users and developers, as well as live integrations with dynamic data.
Thanks for coming along for the ride of this post. ARK has been out here, it's out here, and it's going to continue to be out here, doing its part to make sure everybody knows that blockchains are, in fact, a thing.
BSV Testnet Explorer Auto refreshed every 30 secs. Network Summary. Hashrate . 127.329 TH/s. Unconfirmed Transactions. 2 tx (4.06 KB) Difficulty. 17.29 x 10 6. Exchange Rate -Chainwork. 3.44 x 10 21 hashes. Tools. Node Status. Summary of this node: version, network, uptime, etc. Peers. Detailed info about the peers connected to this node. Browse Blocks. Browse all blocks in the blockchain ... Yes, in bitcoin an address begins with "1" or "3", in testnet bitcoin an address begins with "m" or "2". In the case of Ethereum both addresses can be used, but you can't transfer ether from Rinkeby to the real network. The same for Eos, you can have a testnet account only in testnet and not in mainnet and viceversa. 0000000000000023a280dac57dfe433c57a16a5f1e55ddd8ca8fdc87af6429d0 Time: 2020-10-28 09:33:24 BTC Address. Send coins back, when you don't need them anymore: tb1qm5tfegjevj27yvvna9elym9lnzcf0zraxgl8z2 Height Age Transactions Total Sent Total Fees Block Size (in bytes) 654077: 2020-10-24T07:54:33.84Z: 1,769: 2,703.104 BTC: 0.251 BTC: 862,610: 654076: 2020-10-24T07 ...
I Lived Using Only BITCOIN for 24 hours... - YouTube
L'implementazione Lighning Network su Bitcoin è disponibile sulla rete pubblica testnet, ora si possono avere transazioni economiche e veloci e pagarci il caffé! Vediamo come funziona ... - show you an Ethereum testnet block explorer - what a Bitcoin testnet address looks like - what a faucet is In this video series different topics will be explained which will help you to ... I Survived Living Off Only BITCOIN for 24 hours and it didn't exactly go as expected. Is it possible to live for a day using cryptocurrency? Inspired from Ye... This video is unavailable. Watch Queue Queue. Watch Queue Queue Ravencoin Testnet Faucet and Block Explorer https://testnet.ravencoin.network/ Nothing on this channel is financial advice. My Mining Rig Parts Motherboard: https://amzn.to/2rwd5pf Processor ...