Best mining GPU 2020: the best graphics cards for mining ...

Filecoin | Development Status and Mining Progress

Author: Gamals Ahmed, CoinEx Business Ambassador
https://preview.redd.it/5bqakdqgl3g51.jpg?width=865&format=pjpg&auto=webp&s=b709794863977eb6554e3919b9e00ca750e3e704
A decentralized storage network that transforms cloud storage into an account market. Miners obtain the integrity of the original protocol by providing data storage and / or retrieval. On the contrary, customers pay miners to store or distribute data and retrieve it.
Filecoin announced, that there will be more delays before its main network is officially launched.
Filecoin developers postponed the release date of their main network to late July to late August 2020.
As mentioned in a recent announcement, the Filecoin team said that the initiative completed the first round of the internal protocol security audit. Platform developers claim that the results of the review showed that they need to make several changes to the protocol’s code base before performing the second stage of the software testing process.
Created by Protocol Labs, Filecoin was developed using File System (IPFS), which is a peer-to-peer data storage network. Filecoin will allow users to trade storage space in an open and decentralized market.
Filecoin developers implemented one of the largest cryptocurrency sales in 2017. They have privately obtained over $ 200 million from professional or accredited investors, including many institutional investors.
The main network was slated to launch last month, but in February 2020, the Philly Queen development team delayed the release of the main network between July 15 and July 17, 2020.
They claimed that the outbreak of the Coronavirus (COVID-19) in China was the main cause of the delay. The developers now say that they need more time to solve the problems found during a recent codecase audit.
The Filecoin team noted the following:
“We have drafted a number of protocol changes to ensure that building our major network launch is safe and economically sound.” The project developers will add them to two different implementations of Filecoin (Lotus and go-filecoin) in the coming weeks.
Filecoin developers conducted a survey to allow platform community members to cast their votes on three different launch dates for Testnet Phase 2 and mainnet.
The team reported that the community gave their votes. Based on the vote results, the Filecoin team announced a “conservative” estimate that the second phase of the network test should begin by May 11, 2020. The main Filecoin network may be launched sometime between July 20 and August 21, 2020.
The updates to the project can be found on the Filecoin Road Map.
Filecoin developers stated:
“This option will make us get the most important protocol changes first, and then implement the rest as protocol updates during testnet.” Filecoin is back down from the final test stage.
Another filecoin decentralized storage network provider launched its catalytic test network, the final stage of the storage network test that supports the blockchain.
In a blog post on her website, Filecoin said she will postpone the last test round until August. The company also announced a calibration period from July 20 to August 3 to allow miners to test their mining settings and get an idea of how competition conditions affected their rewards.
Filecoin had announced earlier last month that the catalytic testnet test would precede its flagship launch. The delay in the final test also means that the company has returned the main launch window between August 31 and September 21.
Despite the lack of clear incentives for miners and multiple delays, Filecoin has succeeded in attracting huge interest, especially in China. Investors remained highly speculating on the network’s mining hardware and its premium price.
Mining in Filecoin
In most blockchain protocols, “miners” are network participants who do the work necessary to promote and maintain the blockchain. To provide these services, miners are compensated in the original cryptocurrency.
Mining in Filecoin works completely differently — instead of contributing to computational power, miners contribute storage capacity to use for dealing with customers looking to store data.
Filecoin will contain several types of miners:
Storage miners responsible for storing files and data on the network. Miners retrieval, responsible for providing quick tubes for file recovery. Miners repair to be carried out.
Storage miners are the heart of the network. They earn Filecoin by storing data for clients, and computerizing cipher directories to check storage over time. The probability of earning the reward reward and transaction fees is proportional to the amount of storage that the Miner contributes to the Filecoin network, not the hash power.
Retriever miners are the veins of the network. They earn Filecoin by winning bids and mining fees for a specific file, which is determined by the market value of the said file size. Miners bandwidth and recovery / initial transaction response time will determine its ability to close recovery deals on the network.
The maximum bandwidth of the recovery miners will determine the total amount of deals that it can enter into.
In the current implementation, the focus is mostly on storage miners, who sell storage capacity for FIL.

Hardware recommendations

The current system specifications recommended for running the miner are:
Compared to the hardware requirements for running a validity checker, these standards are much higher — although they definitely deserve it. Since these will not increase in the presumed future, the money spent on Filecoin mining hardware will provide users with many years of reliable service, and they pay themselves many times. Think of investing as a small business for cloud storage. To launch a model on the current data hosting model, it will cost millions of dollars in infrastructure and logistics to get started. With Filecoin, you can do the same for a few thousand dollars.
Proceed to mining
Deals are the primary function of the Filecoin network, and it represents an agreement between a client and miners for a “storage” contract.
Once the customer decides to have a miner to store based on the available capacity, duration and price required, he secures sufficient funds in a linked portfolio to cover the total cost of the deal. The deal is then published once the mine accepts the storage agreement. By default, all Filecoin miners are set to automatically accept any deal that meets their criteria, although this can be disabled for miners who prefer to organize their deals manually.
After the deal is published, the customer prepares the data for storage and then transfers it to the miner. Upon receiving all the data, the miner fills in the data in a sector, closes it, and begins to provide proofs to the chain. Once the first confirmation is obtained, the customer can make sure the data is stored correctly, and the deal has officially started.
Throughout the deal, the miner provides continuous proofs to the chain. Clients gradually pay with money they previously closed. If there is missing or late evidence, the miner is punished. More information about this can be found in the Runtime, Cut and Penalties section of this page.
At Filecoin, miners earn two different types of rewards for their efforts: storage fees and reward prevention.
Storage fees are the fees that customers pay regularly after reaching a deal, in exchange for storing data. This fee is automatically deposited into the withdrawal portfolio associated with miners while they continue to perform their duties over time, and is locked for a short period upon receipt.
Block rewards are large sums given to miners calculated on a new block. Unlike storage fees, these rewards do not come from a linked customer; Instead, the new FIL “prints” the network as an inflationary and incentive measure for miners to develop the chain. All active miners on the network have a chance to get a block bonus, their chance to be directly proportional to the amount of storage space that is currently being contributed to the network.
Duration of operation, cutting and penalties
“Slashing” is a feature found in most blockchain protocols, and is used to punish miners who fail to provide reliable uptime or act maliciously against the network.
In Filecoin, miners are susceptible to two different types of cut: storage error cut, unanimously reduce error.
Storage Error Reduction is a term used to include a wider range of penalties, including error fees, sector penalties, and termination fees. Miners must pay these penalties if they fail to provide reliability of the sector or decide to leave the network voluntarily.
An error fee is a penalty that a miner incurs for each non-working day. Sector punishment: A penalty incurred by a miner of a disrupted sector for which no error was reported before the WindowPoSt inspection.
The sector will pay an error fee after the penalty of the sector once the error is discovered.
Termination Fee: A penalty that a miner incurs when a sector is voluntary or involuntarily terminated and removed from the network.
Cutting consensus error is the penalty that a miner incurs for committing consensus errors. This punishment applies to miners who have acted maliciously against the network consensus function.
Filecoin miners
Eight of the top 10 Felticoin miners are Chinese investors or companies, according to the blockchain explorer, while more companies are selling cloud mining contracts and distributed file sharing system hardware. CoinDesk’s Wolfe Chao wrote: “China’s craze for Filecoin may have been largely related to the long-standing popularity of crypto mining in the country overall, which is home to about 65% of the computing power on Bitcoin at discretion.”
With Filecoin approaching the launch of the mainnet blocknet — after several delays since the $ 200 million increase in 2017 — Chinese investors are once again speculating strongly about network mining devices and their premium prices.
Since Protocol Labs, the company behind Filecoin, released its “Test Incentives” program on June 9 that was scheduled to start in a week’s time, more than a dozen Chinese companies have started selling cloud mining contracts and hardware — despite important details such as economics Mining incentives on the main network are still endless.
Sales volumes to date for each of these companies can range from half a million to tens of millions of dollars, according to self-reported data on these platforms that CoinDesk has watched and interviews with several mining hardware manufacturers.
Filecoin’s goal is to build a distributed storage network with token rewards to spur storage hosting as a way to drive wider adoption. Protocol Labs launched a test network in December 2019. But the tokens mined in the testing environment so far are not representative of the true silicon coin that can be traded when the main network is turned on. Moreover, the mining incentive economics on testnet do not represent how final block rewards will be available on the main network.
However, data from Blockecoin’s blocknetin testnet explorers show that eight out of 10 miners with the most effective mining force on testnet are currently Chinese miners.
These eight miners have about 15 petabytes (PB) of effective storage mining power, accounting for more than 85% of the total test of 17.9 petable. For the context, 1 petabyte of hard disk storage = 1000 terabytes (terabytes) = 1 million gigabytes (GB).
Filecoin craze in China may be closely related to the long-standing popularity of crypt mining in the country overall, which is home to about 65% of the computing power on Bitcoin by estimation. In addition, there has been a lot of hype in China about foreign exchange mining since 2018, as companies promote all types of devices when the network is still in development.
“Encryption mining has always been popular in China,” said Andy Tien, co-founder of 1475, one of several mining hardware manufacturers in Philquin supported by prominent Chinese video indicators such as Fenbushi and Hashkey Capital.
“Even though the Velikoyen mining process is more technologically sophisticated, the idea of mining using hard drives instead of specialized machines like Bitcoin ASIC may be a lot easier for retailers to understand,” he said.
Meanwhile, according to Feixiaohao, a Chinese service comparable to CoinMarketCap, nearly 50 Chinese crypto exchanges are often somewhat unknown with some of the more well-known exchanges including Gate.io and Biki — have listed trading pairs for Filecoin currency contracts for USDT.
In bitcoin mining, at the current difficulty level, one segment per second (TH / s) fragmentation rate is expected to generate around 0.000008 BTC within 24 hours. The higher the number of TH / s, the greater the number of bitcoins it should be able to produce proportionately. But in Filecoin, the efficient mining force of miners depends on the amount of data stamped on the hard drive, not the total size of the hard drive.
To close data in the hard drive, the Filecoin miner still needs processing power, i.e. CPU or GPU as well as RAM. More powerful processors with improved software can confine data to the hard drive more quickly, so miners can combine more efficient mining energy faster on a given day.
As of this stage, there appears to be no transparent way at the network level for retail investors to see how much of the purchased hard disk drive was purchased which actually represents an effective mining force.
The U.S.-based Labs Protocol was behind Filecoin’s initial coin offer for 2017, which raised an astonishing $ 200 million.
This was in addition to a $ 50 million increase in private investment supported by notable venture capital projects including Sequoia, Anderson Horowitz and Union Square Ventures. CoinDk’s parent company, CoinDk, has also invested in Protocol Labs.
After rounds of delay, Protocol Protocols said in September 2019 that a testnet launch would be available around December 2019 and the main network would be rolled out in the first quarter of 2020.
The test started as promised, but the main network has been delayed again and is now expected to launch in August 2020. What is Filecoin mining process?
Filecoin mainly consists of three parts: the storage market (the chain), the blockecin Filecoin, and the search market (under the chain). Storage and research market in series and series respectively for security and efficiency. For users, the storage frequency is relatively low, and the security requirements are relatively high, so the storage process is placed on the chain. The retrieval frequency is much higher than the storage frequency when there is a certain amount of data. Given the performance problem in processing data on the chain, the retrieval process under the chain is performed. In order to solve the security issue of payment in the retrieval process, Filecoin adopts the micro-payment strategy. In simple terms, the process is to split the document into several copies, and every time the user gets a portion of the data, the corresponding fee is paid. Types of mines corresponding to Filecoin’s two major markets are miners and warehousers, among whom miners are primarily responsible for storing data and block packages, while miners are primarily responsible for data query. After the stable operation of the major Filecoin network in the future, the mining operator will be introduced, who is the main responsible for data maintenance.
In the initial release of Filecoin, the request matching mechanism was not implemented in the storage market and retrieval market, but the takeover mechanism was adopted. The three main parts of Filecoin correspond to three processes, namely the stored procedure, retrieval process, packaging and reward process. The following figure shows the simplified process and the income of the miners:
The Filecoin mining process is much more complicated, and the important factor in determining the previous mining profit is efficient storage. Effective storage is a key feature that distinguishes Filecoin from other decentralized storage projects. In Filecoin’s EC consensus, effective storage is similar to interest in PoS, which determines the likelihood that a miner will get the right to fill, that is, the proportion of miners effectively stored in the entire network is proportional to final mining revenue.
It is also possible to obtain higher effective storage under the same hardware conditions by improving the mining algorithm. However, the current increase in the number of benefits that can be achieved by improving the algorithm is still unknown.
It seeks to promote mining using Filecoin Discover
Filecoin announced Filecoin Discover — a step to encourage miners to join the Filecoin network. According to the company, Filecoin Discover is “an ever-growing catalog of numerous petabytes of public data covering literature, science, art, and history.” Miners interested in sharing can choose which data sets they want to store, and receive that data on a drive at a cost. In exchange for storing this verified data, miners will earn additional Filecoin above the regular block rewards for storing data. Includes the current catalog of open source data sets; ENCODE, 1000 Genomes, Project Gutenberg, Berkley Self-driving data, more projects, and datasets are added every day.
Ian Darrow, Head of Operations at Filecoin, commented on the announcement:
“Over 2.5 quintillion bytes of data are created every day. This data includes 294 billion emails, 500 million tweets and 64 billion messages on social media. But it is also climatology reports, disease tracking maps, connected vehicle coordinates and much more. It is extremely important that we maintain data that will serve as the backbone for future research and discovery”.
Miners who choose to participate in Filecoin Discover may receive hard drives pre-loaded with verified data, as well as setup and maintenance instructions, depending on the company. The Filecoin team will also host the Slack (fil-Discover-support) channel where miners can learn more.
Filecoin got its fair share of obstacles along the way. Last month Filecoin announced a further delay before its main network was officially launched — after years of raising funds.
In late July QEBR (OTC: QEBR) announced that it had ceded ownership of two subsidiaries in order to focus all of the company’s resources on building blockchain-based mining operations.
The QEBR technology team previously announced that it has proven its system as a Filecoin node valid with CPU, GPU, bandwidth and storage compatibility that meets all IPFS guidelines. The QEBR test system is connected to the main Filecoin blockchain and the already mined filecoin coin has already been tested.
“The disclosure of Sheen Boom and Jihye will allow our team to focus only on the upcoming global launch of Filecoin. QEBR branch, Shenzhen DZD Digital Technology Ltd. (“ DZD “), has a strong background in blockchain development, extraction Data, data acquisition, data processing, data technology research. We strongly believe Filecoin has the potential to be a leading blockchain-based cryptocurrency and will make every effort to make QEBR an important player when Mainecoin mainnet will be launched soon”.
IPFS and Filecoin
Filecoin and IPFS are complementary protocols for storing and sharing data in a decentralized network. While users are not required to use Filecoin and IPFS together, the two combined are working to resolve major failures in the current web infrastructure.
IPFS
It is an open source protocol that allows users to store and transmit verifiable data with each other. IPFS users insist on data on the network by installing it on their own device, to a third-party cloud service (known as Pinning Services), or through community-oriented systems where a group of individual IPFS users share resources to ensure the content stays live.
The lack of an integrated catalytic mechanism is the challenge Filecoin hopes to solve by allowing users to catalyze long-term distributed storage at competitive prices through the storage contract market, while maintaining the efficiency and flexibility that the IPFS network provides.
Using IPFS
In IPFS, the data is hosted by the required data installation nodes. For data to persist while the user node is offline, users must either rely on their other peers to install their data voluntarily or use a central install service to store data.
Peer-to-peer reliance caching data may be a good thing as one or multiple organizations share common files on an internal network, or where strong social contracts can be used to ensure continued hosting and preservation of content in the long run. Most users in an IPFS network use an installation service.
Using Filecoin
The last option is to install your data in a decentralized storage market, such as Filecoin. In Filecoin’s structure, customers make regular small payments to store data when a certain availability, while miners earn those payments by constantly checking the integrity of this data, storing it, and ensuring its quick recovery. This allows users to motivate Filecoin miners to ensure that their content will be live when it is needed, a distinct advantage of relying only on other network users as required using IPFS alone.
Filecoin, powered by IPFS
It is important to know that Filecoin is built on top of IPFS. Filecoin aims to be a very integrated and seamless storage market that takes advantage of the basic functions provided by IPFS, they are connected to each other, but can be implemented completely independently of each other. Users do not need to interact with Filecoin in order to use IPFS.
Some advantages of sharing Filecoin with IPFS:
Of all the decentralized storage projects, Filecoin is undoubtedly the most interested, and IPFS has been running stably for two years, fully demonstrating the strength of its core protocol.
Filecoin’s ability to obtain market share from traditional central storage depends on end-user experience and storage price. Currently, most Filecoin nodes are posted in the IDC room. Actual deployment and operation costs are not reduced compared to traditional central cloud storage, and the storage process is more complicated.
PoRep and PoSt, which has a large number of proofs of unknown operation, are required to cause the actual storage cost to be so, in the early days of the release of Filecoin. The actual cost of storing data may be higher than the cost of central cloud storage, but the initial storage node may reduce the storage price in order to obtain block rewards, which may result in the actual storage price lower than traditional central cloud storage.
In the long term, Filecoin still needs to take full advantage of its P2P storage, convert storage devices from specialization to civil use, and improve its algorithms to reduce storage costs without affecting user experience. The storage problem is an important problem to be solved in the blockchain field, so a large number of storage projects were presented at the 19th Web3 Summit. IPFS is an important part of Web3 visibility. Its development will affect the development of Web3 to some extent. Likewise, Web3 development somewhat determines the future of IPFS. Filecoin is an IPFS-based storage class project initiated by IPFS. There is no doubt that he is highly expected.
Resources :
  1. https://www.coindesk.com/filecoin-pushes-back-final-testing-phase-announces-calibration-period-for-miners
  2. https://docs.filecoin.io/mine/#types-of-miners https://www.nasdaq.com/articles/inside-the-craze-for-filecoin-crypto-mining-in-china-2020-07-12؟amp
  3. https://www.prnewswire.com/news-releases/qebr-streamlines-holdings-to-concentrate-on-filecoin-development-and-mining-301098731.html
  4. https://www.crowdfundinsider.com/2020/05/161200-filecoin-seeks-to-boost-mining-with-filecoin-discove
  5. https://zephyrnet.com/filecoin-seeks-to-boost-mining-with-filecoin-discove
  6. https://docs.filecoin.io/introduction/ipfs-and-filecoin/#filecoin-powered-by-ipfs
submitted by CoinEx_Institution to filecoin [link] [comments]

Cryptocurrency Mining Today

Cryptocurrency Mining Today
Mining is one of the key concepts in the crypto world. Everyone who comes into contact with this sphere somehow wonders about the mining of coins. How profitable is mining in 2020, and what are the current trends?
by StealthEX
Crypto mining is a process during which a computer solves mathematical problems, resulting in the release of new blocks of information. This gives its owners a certain amount of coins, which is deposited in the total pot and registered in the public “ledger”, so-called blockchain. Machines in the network are also checking transactions with existing coins, adding this information to the blockchain as well.
As for the issue itself, the most well-known algorithm of mining is Proof-of-Work (PoW), used in the networks of Bitcoin, Litecoin, Ethereum and many others.
During the mining process, the latest transactions are verified and compiled into blocks. It is usually a series of calculations with an iteration of parameters to find a hash with the specified properties. The node which first solves this problem receives a reward. This approach was specifically designed to encourage those who provide the computing power of their mining machines to maintain the network and mine new coins.
It is usually no need for a newcomer to know and understand all the complicated details of the mining process, just how much they can earn with certain equipment and electricity costs.
Everything is designed in such a way that the complexity of calculations is steadily increasing, which then requires a constant increase in the computing power of the network. In 2009-2010, for mining bitcoin, miners only had to download and run the software on their personal computers, but very soon the network became so complicated that even with best PCs with a powerful processor, mining became unprofitable. That’s why miners started to use more effective video cards (graphics processing units or GPUs) and join them in so-called “farms”.
In most systems, the number of coins is determined in advance. Also, many networks are gradually reducing rewards for miners. Such emission restrictions were built into the algorithm to prevent inflation.
Thus, the cost of mining for smaller participants no longer pays off, which makes them turn off their hardware or switch to another coin where they can still make their profit.
In particular, on the evening of May 11, 2020, a halving took place in the bitcoin network, the reward for mining was halved, from 12.5 to 6.25 BTC. In June, the revenue of bitcoin miners decreased by 23%, to the lowest since March 2019.
However, in mid-June, the difficulty of bitcoin mining showed a record growth over the past 2.5 years. Mining the first cryptocurrency has become 15% more difficult. Although, by the beginning of July, the complexity had stabilized. The growing difficulty of mining the first cryptocurrency indicates that new miners have joined its network. Previously, some of them turned off the equipment, as it became less profitable to mine the coin due to a decrease in its cost and halving.
Now the absolute majority of new coins are generated by industrial mining. This is done by large data centers equipped with specialized computers based on the ASIC architecture. ASICs are integrated circuits that were initially optimized for a specific task, namely the mining of cryptocurrencies. They are much more productive than CPUs and video cards, and at the same time consume much less electricity. ASIC computers are the main type of equipment for the industrial production of crypto.
So now, after the halving, BTC coin mining has become even less profitable. For beginners, mining the first cryptocurrency is unlikely to be suitable. It is more often earned by large companies that have all the necessary equipment, access to cheap rental conditions, electricity and maintenance.
Hence newbies are better off starting with mining altcoins. It is even more profitable to work in a pool, that is, together with other miners. This can help to place farms in one place and negotiate a favourable price for electricity, so you can get a small but stable income dux to the total capacity of the pool.
Therefore, it has become much more difficult for regular users who have only non-specialized equipment at their disposal to generate virtual money. However, GPU developers have significantly increased the performance of their devices in recent years, so mining on a video card is still common.
Another important event that changes the situation in the mining sphere will be the hardfork of the Ethereum network with the turn to the Proof-of-Stake algorithm. For now, Ethereum is the most popular altcoin for GPU mining, but Ethereum 2.0 will not require using such powerful equipment, so then it switches to PoS, GPU owners will have to look for alternative coins to mine.
At the moment the most popular altcoins for mining on GPUs are Ethereum (ETH), Ethereum Classic (ETC), Grin (GRIN), Zcoin (XZC), Dogecoin and Ravencoin (RVN). There are actually a lot of mining programs that automatically determine which coin is more profitable to mine at the moment.
In the coming years, the market is waiting for a race of technologies. Manufacturers are investing in finding ways to increase hashing speed and reduce power consumption. Mining pools will play an increasing role. The market will also be affected by applications for mining cryptocurrencies on smartphones that require low computing power, such as Dash or Litecoin.
And remember StealthEX supports more than 250 coins and constantly updating the list, so you can easily swap your crypto haul to more popular altcoins. Our service does not require registration and allows you to remain anonymous. Why don’t you check it out? Just go to StealthEX and follow these easy steps:
✔ Choose the pair and the amount for your exchange. For example ETH to BTC.
✔ Press the “Start exchange” button.
✔ Provide the recipient address to which the coins will be transferred.
✔ Move your cryptocurrency for the exchange.
✔ Receive your coins.
Follow us on Medium, Twitter, and Reddit to get StealthEX.io updates and the latest news about the crypto world. For all requests message us via [email protected].
The views and opinions expressed here are solely those of the author. Every investment and trading move involves risk. You should conduct your own research when making a decision.
Original article was posted on https://stealthex.io/blog/2020/07/28/mining-today/
submitted by Stealthex_io to StealthEX [link] [comments]

Why i’m bullish on Zilliqa (long read)

Edit: TL;DR added in the comments
 
Hey all, I've been researching coins since 2017 and have gone through 100s of them in the last 3 years. I got introduced to blockchain via Bitcoin of course, analyzed Ethereum thereafter and from that moment I have a keen interest in smart contact platforms. I’m passionate about Ethereum but I find Zilliqa to have a better risk-reward ratio. Especially because Zilliqa has found an elegant balance between being secure, decentralized and scalable in my opinion.
 
Below I post my analysis of why from all the coins I went through I’m most bullish on Zilliqa (yes I went through Tezos, EOS, NEO, VeChain, Harmony, Algorand, Cardano etc.). Note that this is not investment advice and although it's a thorough analysis there is obviously some bias involved. Looking forward to what you all think!
 
Fun fact: the name Zilliqa is a play on ‘silica’ silicon dioxide which means “Silicon for the high-throughput consensus computer.”
 
This post is divided into (i) Technology, (ii) Business & Partnerships, and (iii) Marketing & Community. I’ve tried to make the technology part readable for a broad audience. If you’ve ever tried understanding the inner workings of Bitcoin and Ethereum you should be able to grasp most parts. Otherwise, just skim through and once you are zoning out head to the next part.
 
Technology and some more:
 
Introduction
 
The technology is one of the main reasons why I’m so bullish on Zilliqa. First thing you see on their website is: “Zilliqa is a high-performance, high-security blockchain platform for enterprises and next-generation applications.” These are some bold statements.
 
Before we deep dive into the technology let’s take a step back in time first as they have quite the history. The initial research paper from which Zilliqa originated dates back to August 2016: Elastico: A Secure Sharding Protocol For Open Blockchains where Loi Luu (Kyber Network) is one of the co-authors. Other ideas that led to the development of what Zilliqa has become today are: Bitcoin-NG, collective signing CoSi, ByzCoin and Omniledger.
 
The technical white paper was made public in August 2017 and since then they have achieved everything stated in the white paper and also created their own open source intermediate level smart contract language called Scilla (functional programming language similar to OCaml) too.
 
Mainnet is live since the end of January 2019 with daily transaction rates growing continuously. About a week ago mainnet reached 5 million transactions, 500.000+ addresses in total along with 2400 nodes keeping the network decentralized and secure. Circulating supply is nearing 11 billion and currently only mining rewards are left. The maximum supply is 21 billion with annual inflation being 7.13% currently and will only decrease with time.
 
Zilliqa realized early on that the usage of public cryptocurrencies and smart contracts were increasing but decentralized, secure, and scalable alternatives were lacking in the crypto space. They proposed to apply sharding onto a public smart contract blockchain where the transaction rate increases almost linear with the increase in the amount of nodes. More nodes = higher transaction throughput and increased decentralization. Sharding comes in many forms and Zilliqa uses network-, transaction- and computational sharding. Network sharding opens up the possibility of using transaction- and computational sharding on top. Zilliqa does not use state sharding for now. We’ll come back to this later.
 
Before we continue dissecting how Zilliqa achieves such from a technological standpoint it’s good to keep in mind that a blockchain being decentralised and secure and scalable is still one of the main hurdles in allowing widespread usage of decentralised networks. In my opinion this needs to be solved first before blockchains can get to the point where they can create and add large scale value. So I invite you to read the next section to grasp the underlying fundamentals. Because after all these premises need to be true otherwise there isn’t a fundamental case to be bullish on Zilliqa, right?
 
Down the rabbit hole
 
How have they achieved this? Let’s define the basics first: key players on Zilliqa are the users and the miners. A user is anybody who uses the blockchain to transfer funds or run smart contracts. Miners are the (shard) nodes in the network who run the consensus protocol and get rewarded for their service in Zillings (ZIL). The mining network is divided into several smaller networks called shards, which is also referred to as ‘network sharding’. Miners subsequently are randomly assigned to a shard by another set of miners called DS (Directory Service) nodes. The regular shards process transactions and the outputs of these shards are eventually combined by the DS shard as they reach consensus on the final state. More on how these DS shards reach consensus (via pBFT) will be explained later on.
 
The Zilliqa network produces two types of blocks: DS blocks and Tx blocks. One DS Block consists of 100 Tx Blocks. And as previously mentioned there are two types of nodes concerned with reaching consensus: shard nodes and DS nodes. Becoming a shard node or DS node is being defined by the result of a PoW cycle (Ethash) at the beginning of the DS Block. All candidate mining nodes compete with each other and run the PoW (Proof-of-Work) cycle for 60 seconds and the submissions achieving the highest difficulty will be allowed on the network. And to put it in perspective: the average difficulty for one DS node is ~ 2 Th/s equaling 2.000.000 Mh/s or 55 thousand+ GeForce GTX 1070 / 8 GB GPUs at 35.4 Mh/s. Each DS Block 10 new DS nodes are allowed. And a shard node needs to provide around 8.53 GH/s currently (around 240 GTX 1070s). Dual mining ETH/ETC and ZIL is possible and can be done via mining software such as Phoenix and Claymore. There are pools and if you have large amounts of hashing power (Ethash) available you could mine solo.
 
The PoW cycle of 60 seconds is a peak performance and acts as an entry ticket to the network. The entry ticket is called a sybil resistance mechanism and makes it incredibly hard for adversaries to spawn lots of identities and manipulate the network with these identities. And after every 100 Tx Blocks which corresponds to roughly 1,5 hour this PoW process repeats. In between these 1,5 hour, no PoW needs to be done meaning Zilliqa’s energy consumption to keep the network secure is low. For more detailed information on how mining works click here.
Okay, hats off to you. You have made it this far. Before we go any deeper down the rabbit hole we first must understand why Zilliqa goes through all of the above technicalities and understand a bit more what a blockchain on a more fundamental level is. Because the core of Zilliqa’s consensus protocol relies on the usage of pBFT (practical Byzantine Fault Tolerance) we need to know more about state machines and their function. Navigate to Viewblock, a Zilliqa block explorer, and just come back to this article. We will use this site to navigate through a few concepts.
 
We have established that Zilliqa is a public and distributed blockchain. Meaning that everyone with an internet connection can send ZILs, trigger smart contracts, etc. and there is no central authority who fully controls the network. Zilliqa and other public and distributed blockchains (like Bitcoin and Ethereum) can also be defined as state machines.
 
Taking the liberty of paraphrasing examples and definitions given by Samuel Brooks’ medium article, he describes the definition of a blockchain (like Zilliqa) as: “A peer-to-peer, append-only datastore that uses consensus to synchronize cryptographically-secure data”.
 
Next, he states that: "blockchains are fundamentally systems for managing valid state transitions”. For some more context, I recommend reading the whole medium article to get a better grasp of the definitions and understanding of state machines. Nevertheless, let’s try to simplify and compile it into a single paragraph. Take traffic lights as an example: all its states (red, amber, and green) are predefined, all possible outcomes are known and it doesn’t matter if you encounter the traffic light today or tomorrow. It will still behave the same. Managing the states of a traffic light can be done by triggering a sensor on the road or pushing a button resulting in one traffic lights’ state going from green to red (via amber) and another light from red to green.
 
With public blockchains like Zilliqa, this isn’t so straightforward and simple. It started with block #1 almost 1,5 years ago and every 45 seconds or so a new block linked to the previous block is being added. Resulting in a chain of blocks with transactions in it that everyone can verify from block #1 to the current #647.000+ block. The state is ever changing and the states it can find itself in are infinite. And while the traffic light might work together in tandem with various other traffic lights, it’s rather insignificant comparing it to a public blockchain. Because Zilliqa consists of 2400 nodes who need to work together to achieve consensus on what the latest valid state is while some of these nodes may have latency or broadcast issues, drop offline or are deliberately trying to attack the network, etc.
 
Now go back to the Viewblock page take a look at the amount of transaction, addresses, block and DS height and then hit refresh. Obviously as expected you see new incremented values on one or all parameters. And how did the Zilliqa blockchain manage to transition from a previous valid state to the latest valid state? By using pBFT to reach consensus on the latest valid state.
 
After having obtained the entry ticket, miners execute pBFT to reach consensus on the ever-changing state of the blockchain. pBFT requires a series of network communication between nodes, and as such there is no GPU involved (but CPU). Resulting in the total energy consumed to keep the blockchain secure, decentralized and scalable being low.
 
pBFT stands for practical Byzantine Fault Tolerance and is an optimization on the Byzantine Fault Tolerant algorithm. To quote Blockonomi: “In the context of distributed systems, Byzantine Fault Tolerance is the ability of a distributed computer network to function as desired and correctly reach a sufficient consensus despite malicious components (nodes) of the system failing or propagating incorrect information to other peers.” Zilliqa is such a distributed computer network and depends on the honesty of the nodes (shard and DS) to reach consensus and to continuously update the state with the latest block. If pBFT is a new term for you I can highly recommend the Blockonomi article.
 
The idea of pBFT was introduced in 1999 - one of the authors even won a Turing award for it - and it is well researched and applied in various blockchains and distributed systems nowadays. If you want more advanced information than the Blockonomi link provides click here. And if you’re in between Blockonomi and the University of Singapore read the Zilliqa Design Story Part 2 dating from October 2017.
Quoting from the Zilliqa tech whitepaper: “pBFT relies upon a correct leader (which is randomly selected) to begin each phase and proceed when the sufficient majority exists. In case the leader is byzantine it can stall the entire consensus protocol. To address this challenge, pBFT offers a view change protocol to replace the byzantine leader with another one.”
 
pBFT can tolerate ⅓ of the nodes being dishonest (offline counts as Byzantine = dishonest) and the consensus protocol will function without stalling or hiccups. Once there are more than ⅓ of dishonest nodes but no more than ⅔ the network will be stalled and a view change will be triggered to elect a new DS leader. Only when more than ⅔ of the nodes are dishonest (66%) double-spend attacks become possible.
 
If the network stalls no transactions can be processed and one has to wait until a new honest leader has been elected. When the mainnet was just launched and in its early phases, view changes happened regularly. As of today the last stalling of the network - and view change being triggered - was at the end of October 2019.
 
Another benefit of using pBFT for consensus besides low energy is the immediate finality it provides. Once your transaction is included in a block and the block is added to the chain it’s done. Lastly, take a look at this article where three types of finality are being defined: probabilistic, absolute and economic finality. Zilliqa falls under the absolute finality (just like Tendermint for example). Although lengthy already we skipped through some of the inner workings from Zilliqa’s consensus: read the Zilliqa Design Story Part 3 and you will be close to having a complete picture on it. Enough about PoW, sybil resistance mechanism, pBFT, etc. Another thing we haven’t looked at yet is the amount of decentralization.
 
Decentralisation
 
Currently, there are four shards, each one of them consisting of 600 nodes. 1 shard with 600 so-called DS nodes (Directory Service - they need to achieve a higher difficulty than shard nodes) and 1800 shard nodes of which 250 are shard guards (centralized nodes controlled by the team). The amount of shard guards has been steadily declining from 1200 in January 2019 to 250 as of May 2020. On the Viewblock statistics, you can see that many of the nodes are being located in the US but those are only the (CPU parts of the) shard nodes who perform pBFT. There is no data from where the PoW sources are coming. And when the Zilliqa blockchain starts reaching its transaction capacity limit, a network upgrade needs to be executed to lift the current cap of maximum 2400 nodes to allow more nodes and formation of more shards which will allow to network to keep on scaling according to demand.
Besides shard nodes there are also seed nodes. The main role of seed nodes is to serve as direct access points (for end-users and clients) to the core Zilliqa network that validates transactions. Seed nodes consolidate transaction requests and forward these to the lookup nodes (another type of nodes) for distribution to the shards in the network. Seed nodes also maintain the entire transaction history and the global state of the blockchain which is needed to provide services such as block explorers. Seed nodes in the Zilliqa network are comparable to Infura on Ethereum.
 
The seed nodes were first only operated by Zilliqa themselves, exchanges and Viewblock. Operators of seed nodes like exchanges had no incentive to open them for the greater public. They were centralised at first. Decentralisation at the seed nodes level has been steadily rolled out since March 2020 ( Zilliqa Improvement Proposal 3 ). Currently the amount of seed nodes is being increased, they are public-facing and at the same time PoS is applied to incentivize seed node operators and make it possible for ZIL holders to stake and earn passive yields. Important distinction: seed nodes are not involved with consensus! That is still PoW as entry ticket and pBFT for the actual consensus.
 
5% of the block rewards are being assigned to seed nodes (from the beginning in 2019) and those are being used to pay out ZIL stakers. The 5% block rewards with an annual yield of 10.03% translate to roughly 610 MM ZILs in total that can be staked. Exchanges use the custodial variant of staking and wallets like Moonlet will use the non-custodial version (starting in Q3 2020). Staking is being done by sending ZILs to a smart contract created by Zilliqa and audited by Quantstamp.
 
With a high amount of DS; shard nodes and seed nodes becoming more decentralized too, Zilliqa qualifies for the label of decentralized in my opinion.
 
Smart contracts
 
Let me start by saying I’m not a developer and my programming skills are quite limited. So I‘m taking the ELI5 route (maybe 12) but if you are familiar with Javascript, Solidity or specifically OCaml please head straight to Scilla - read the docs to get a good initial grasp of how Zilliqa’s smart contract language Scilla works and if you ask yourself “why another programming language?” check this article. And if you want to play around with some sample contracts in an IDE click here. The faucet can be found here. And more information on architecture, dapp development and API can be found on the Developer Portal.
If you are more into listening and watching: check this recent webinar explaining Zilliqa and Scilla. Link is time-stamped so you’ll start right away with a platform introduction, roadmap 2020 and afterwards a proper Scilla introduction.
 
Generalized: programming languages can be divided into being ‘object-oriented’ or ‘functional’. Here is an ELI5 given by software development academy: * “all programs have two basic components, data – what the program knows – and behavior – what the program can do with that data. So object-oriented programming states that combining data and related behaviors in one place, is called “object”, which makes it easier to understand how a particular program works. On the other hand, functional programming argues that data and behavior are different things and should be separated to ensure their clarity.” *
 
Scilla is on the functional side and shares similarities with OCaml: OCaml is a general-purpose programming language with an emphasis on expressiveness and safety. It has an advanced type system that helps catch your mistakes without getting in your way. It's used in environments where a single mistake can cost millions and speed matters, is supported by an active community, and has a rich set of libraries and development tools. For all its power, OCaml is also pretty simple, which is one reason it's often used as a teaching language.
 
Scilla is blockchain agnostic, can be implemented onto other blockchains as well, is recognized by academics and won a so-called Distinguished Artifact Award award at the end of last year.
 
One of the reasons why the Zilliqa team decided to create their own programming language focused on preventing smart contract vulnerabilities is that adding logic on a blockchain, programming, means that you cannot afford to make mistakes. Otherwise, it could cost you. It’s all great and fun blockchains being immutable but updating your code because you found a bug isn’t the same as with a regular web application for example. And with smart contracts, it inherently involves cryptocurrencies in some form thus value.
 
Another difference with programming languages on a blockchain is gas. Every transaction you do on a smart contract platform like Zilliqa or Ethereum costs gas. With gas you basically pay for computational costs. Sending a ZIL from address A to address B costs 0.001 ZIL currently. Smart contracts are more complex, often involve various functions and require more gas (if gas is a new concept click here ).
 
So with Scilla, similar to Solidity, you need to make sure that “every function in your smart contract will run as expected without hitting gas limits. An improper resource analysis may lead to situations where funds may get stuck simply because a part of the smart contract code cannot be executed due to gas limits. Such constraints are not present in traditional software systems”. Scilla design story part 1
 
Some examples of smart contract issues you’d want to avoid are: leaking funds, ‘unexpected changes to critical state variables’ (example: someone other than you setting his or her address as the owner of the smart contract after creation) or simply killing a contract.
 
Scilla also allows for formal verification. Wikipedia to the rescue: In the context of hardware and software systems, formal verification is the act of proving or disproving the correctness of intended algorithms underlying a system with respect to a certain formal specification or property, using formal methods of mathematics.
 
Formal verification can be helpful in proving the correctness of systems such as: cryptographic protocols, combinational circuits, digital circuits with internal memory, and software expressed as source code.
 
Scilla is being developed hand-in-hand with formalization of its semantics and its embedding into the Coq proof assistant — a state-of-the art tool for mechanized proofs about properties of programs.”
 
Simply put, with Scilla and accompanying tooling developers can be mathematically sure and proof that the smart contract they’ve written does what he or she intends it to do.
 
Smart contract on a sharded environment and state sharding
 
There is one more topic I’d like to touch on: smart contract execution in a sharded environment (and what is the effect of state sharding). This is a complex topic. I’m not able to explain it any easier than what is posted here. But I will try to compress the post into something easy to digest.
 
Earlier on we have established that Zilliqa can process transactions in parallel due to network sharding. This is where the linear scalability comes from. We can define simple transactions: a transaction from address A to B (Category 1), a transaction where a user interacts with one smart contract (Category 2) and the most complex ones where triggering a transaction results in multiple smart contracts being involved (Category 3). The shards are able to process transactions on their own without interference of the other shards. With Category 1 transactions that is doable, with Category 2 transactions sometimes if that address is in the same shard as the smart contract but with Category 3 you definitely need communication between the shards. Solving that requires to make a set of communication rules the protocol needs to follow in order to process all transactions in a generalised fashion.
 
And this is where the downsides of state sharding comes in currently. All shards in Zilliqa have access to the complete state. Yes the state size (0.1 GB at the moment) grows and all of the nodes need to store it but it also means that they don’t need to shop around for information available on other shards. Requiring more communication and adding more complexity. Computer science knowledge and/or developer knowledge required links if you want to dig further: Scilla - language grammar Scilla - Foundations for Verifiable Decentralised Computations on a Blockchain Gas Accounting NUS x Zilliqa: Smart contract language workshop
 
Easier to follow links on programming Scilla https://learnscilla.com/home Ivan on Tech
 
Roadmap / Zilliqa 2.0
 
There is no strict defined roadmap but here are topics being worked on. And via the Zilliqa website there is also more information on the projects they are working on.
 
Business & Partnerships
 
It’s not only technology in which Zilliqa seems to be excelling as their ecosystem has been expanding and starting to grow rapidly. The project is on a mission to provide OpenFinance (OpFi) to the world and Singapore is the right place to be due to its progressive regulations and futuristic thinking. Singapore has taken a proactive approach towards cryptocurrencies by introducing the Payment Services Act 2019 (PS Act). Among other things, the PS Act will regulate intermediaries dealing with certain cryptocurrencies, with a particular focus on consumer protection and anti-money laundering. It will also provide a stable regulatory licensing and operating framework for cryptocurrency entities, effectively covering all crypto businesses and exchanges based in Singapore. According to PWC 82% of the surveyed executives in Singapore reported blockchain initiatives underway and 13% of them have already brought the initiatives live to the market. There is also an increasing list of organizations that are starting to provide digital payment services. Moreover, Singaporean blockchain developers Building Cities Beyond has recently created an innovation $15 million grant to encourage development on its ecosystem. This all suggests that Singapore tries to position itself as (one of) the leading blockchain hubs in the world.
 
Zilliqa seems to already take advantage of this and recently helped launch Hg Exchange on their platform, together with financial institutions PhillipCapital, PrimePartners and Fundnel. Hg Exchange, which is now approved by the Monetary Authority of Singapore (MAS), uses smart contracts to represent digital assets. Through Hg Exchange financial institutions worldwide can use Zilliqa's safe-by-design smart contracts to enable the trading of private equities. For example, think of companies such as Grab, Airbnb, SpaceX that are not available for public trading right now. Hg Exchange will allow investors to buy shares of private companies & unicorns and capture their value before an IPO. Anquan, the main company behind Zilliqa, has also recently announced that they became a partner and shareholder in TEN31 Bank, which is a fully regulated bank allowing for tokenization of assets and is aiming to bridge the gap between conventional banking and the blockchain world. If STOs, the tokenization of assets, and equity trading will continue to increase, then Zilliqa’s public blockchain would be the ideal candidate due to its strategic positioning, partnerships, regulatory compliance and the technology that is being built on top of it.
 
What is also very encouraging is their focus on banking the un(der)banked. They are launching a stablecoin basket starting with XSGD. As many of you know, stablecoins are currently mostly used for trading. However, Zilliqa is actively trying to broaden the use case of stablecoins. I recommend everybody to read this text that Amrit Kumar wrote (one of the co-founders). These stablecoins will be integrated in the traditional markets and bridge the gap between the crypto world and the traditional world. This could potentially revolutionize and legitimise the crypto space if retailers and companies will for example start to use stablecoins for payments or remittances, instead of it solely being used for trading.
 
Zilliqa also released their DeFi strategic roadmap (dating November 2019) which seems to be aligning well with their OpFi strategy. A non-custodial DEX is coming to Zilliqa made by Switcheo which allows cross-chain trading (atomic swaps) between ETH, EOS and ZIL based tokens. They also signed a Memorandum of Understanding for a (soon to be announced) USD stablecoin. And as Zilliqa is all about regulations and being compliant, I’m speculating on it to be a regulated USD stablecoin. Furthermore, XSGD is already created and visible on block explorer and XIDR (Indonesian Stablecoin) is also coming soon via StraitsX. Here also an overview of the Tech Stack for Financial Applications from September 2019. Further quoting Amrit Kumar on this:
 
There are two basic building blocks in DeFi/OpFi though: 1) stablecoins as you need a non-volatile currency to get access to this market and 2) a dex to be able to trade all these financial assets. The rest are built on top of these blocks.
 
So far, together with our partners and community, we have worked on developing these building blocks with XSGD as a stablecoin. We are working on bringing a USD-backed stablecoin as well. We will soon have a decentralised exchange developed by Switcheo. And with HGX going live, we are also venturing into the tokenization space. More to come in the future.”
 
Additionally, they also have this ZILHive initiative that injects capital into projects. There have been already 6 waves of various teams working on infrastructure, innovation and research, and they are not from ASEAN or Singapore only but global: see Grantees breakdown by country. Over 60 project teams from over 20 countries have contributed to Zilliqa's ecosystem. This includes individuals and teams developing wallets, explorers, developer toolkits, smart contract testing frameworks, dapps, etc. As some of you may know, Unstoppable Domains (UD) blew up when they launched on Zilliqa. UD aims to replace cryptocurrency addresses with a human-readable name and allows for uncensorable websites. Zilliqa will probably be the only one able to handle all these transactions onchain due to ability to scale and its resulting low fees which is why the UD team launched this on Zilliqa in the first place. Furthermore, Zilliqa also has a strong emphasis on security, compliance, and privacy, which is why they partnered with companies like Elliptic, ChainSecurity (part of PwC Switzerland), and Incognito. Their sister company Aqilliz (Zilliqa spelled backwards) focuses on revolutionizing the digital advertising space and is doing interesting things like using Zilliqa to track outdoor digital ads with companies like Foodpanda.
 
Zilliqa is listed on nearly all major exchanges, having several different fiat-gateways and recently have been added to Binance’s margin trading and futures trading with really good volume. They also have a very impressive team with good credentials and experience. They don't just have “tech people”. They have a mix of tech people, business people, marketeers, scientists, and more. Naturally, it's good to have a mix of people with different skill sets if you work in the crypto space.
 
Marketing & Community
 
Zilliqa has a very strong community. If you just follow their Twitter their engagement is much higher for a coin that has approximately 80k followers. They also have been ‘coin of the day’ by LunarCrush many times. LunarCrush tracks real-time cryptocurrency value and social data. According to their data, it seems Zilliqa has a more fundamental and deeper understanding of marketing and community engagement than almost all other coins. While almost all coins have been a bit frozen in the last months, Zilliqa seems to be on its own bull run. It was somewhere in the 100s a few months ago and is currently ranked #46 on CoinGecko. Their official Telegram also has over 20k people and is very active, and their community channel which is over 7k now is more active and larger than many other official channels. Their local communities also seem to be growing.
 
Moreover, their community started ‘Zillacracy’ together with the Zilliqa core team ( see www.zillacracy.com ). It’s a community-run initiative where people from all over the world are now helping with marketing and development on Zilliqa. Since its launch in February 2020 they have been doing a lot and will also run their own non-custodial seed node for staking. This seed node will also allow them to start generating revenue for them to become a self sustaining entity that could potentially scale up to become a decentralized company working in parallel with the Zilliqa core team. Comparing it to all the other smart contract platforms (e.g. Cardano, EOS, Tezos etc.) they don't seem to have started a similar initiative (correct me if I’m wrong though). This suggests in my opinion that these other smart contract platforms do not fully understand how to utilize the ‘power of the community’. This is something you cannot ‘buy with money’ and gives many projects in the space a disadvantage.
 
Zilliqa also released two social products called SocialPay and Zeeves. SocialPay allows users to earn ZILs while tweeting with a specific hashtag. They have recently used it in partnership with the Singapore Red Cross for a marketing campaign after their initial pilot program. It seems like a very valuable social product with a good use case. I can see a lot of traditional companies entering the space through this product, which they seem to suggest will happen. Tokenizing hashtags with smart contracts to get network effect is a very smart and innovative idea.
 
Regarding Zeeves, this is a tipping bot for Telegram. They already have 1000s of signups and they plan to keep upgrading it for more and more people to use it (e.g. they recently have added a quiz features). They also use it during AMAs to reward people in real-time. It’s a very smart approach to grow their communities and get familiar with ZIL. I can see this becoming very big on Telegram. This tool suggests, again, that the Zilliqa team has a deeper understanding of what the crypto space and community needs and is good at finding the right innovative tools to grow and scale.
 
To be honest, I haven’t covered everything (i’m also reaching the character limited haha). So many updates happening lately that it's hard to keep up, such as the International Monetary Fund mentioning Zilliqa in their report, custodial and non-custodial Staking, Binance Margin, Futures, Widget, entering the Indian market, and more. The Head of Marketing Colin Miles has also released this as an overview of what is coming next. And last but not least, Vitalik Buterin has been mentioning Zilliqa lately acknowledging Zilliqa and mentioning that both projects have a lot of room to grow. There is much more info of course and a good part of it has been served to you on a silver platter. I invite you to continue researching by yourself :-) And if you have any comments or questions please post here!
submitted by haveyouheardaboutit to CryptoCurrency [link] [comments]

Why i’m bullish on Zilliqa (long read)

Hey all, I've been researching coins since 2017 and have gone through 100s of them in the last 3 years. I got introduced to blockchain via Bitcoin of course, analysed Ethereum thereafter and from that moment I have a keen interest in smart contact platforms. I’m passionate about Ethereum but I find Zilliqa to have a better risk reward ratio. Especially because Zilliqa has found an elegant balance between being secure, decentralised and scalable in my opinion.
 
Below I post my analysis why from all the coins I went through I’m most bullish on Zilliqa (yes I went through Tezos, EOS, NEO, VeChain, Harmony, Algorand, Cardano etc.). Note that this is not investment advice and although it's a thorough analysis there is obviously some bias involved. Looking forward to what you all think!
 
Fun fact: the name Zilliqa is a play on ‘silica’ silicon dioxide which means “Silicon for the high-throughput consensus computer.”
 
This post is divided into (i) Technology, (ii) Business & Partnerships, and (iii) Marketing & Community. I’ve tried to make the technology part readable for a broad audience. If you’ve ever tried understanding the inner workings of Bitcoin and Ethereum you should be able to grasp most parts. Otherwise just skim through and once you are zoning out head to the next part.
 
Technology and some more:
 
Introduction The technology is one of the main reasons why I’m so bullish on Zilliqa. First thing you see on their website is: “Zilliqa is a high-performance, high-security blockchain platform for enterprises and next-generation applications.” These are some bold statements.
 
Before we deep dive into the technology let’s take a step back in time first as they have quite the history. The initial research paper from which Zilliqa originated dates back to August 2016: Elastico: A Secure Sharding Protocol For Open Blockchains where Loi Luu (Kyber Network) is one of the co-authors. Other ideas that led to the development of what Zilliqa has become today are: Bitcoin-NG, collective signing CoSi, ByzCoin and Omniledger.
 
The technical white paper was made public in August 2017 and since then they have achieved everything stated in the white paper and also created their own open source intermediate level smart contract language called Scilla (functional programming language similar to OCaml) too.
 
Mainnet is live since end of January 2019 with daily transaction rate growing continuously. About a week ago mainnet reached 5 million transactions, 500.000+ addresses in total along with 2400 nodes keeping the network decentralised and secure. Circulating supply is nearing 11 billion and currently only mining rewards are left. Maximum supply is 21 billion with annual inflation being 7.13% currently and will only decrease with time.
 
Zilliqa realised early on that the usage of public cryptocurrencies and smart contracts were increasing but decentralised, secure and scalable alternatives were lacking in the crypto space. They proposed to apply sharding onto a public smart contract blockchain where the transaction rate increases almost linear with the increase in amount of nodes. More nodes = higher transaction throughput and increased decentralisation. Sharding comes in many forms and Zilliqa uses network-, transaction- and computational sharding. Network sharding opens up the possibility of using transaction- and computational sharding on top. Zilliqa does not use state sharding for now. We’ll come back to this later.
 
Before we continue disecting how Zilliqa achieves such from a technological standpoint it’s good to keep in mind that a blockchain being decentralised and secure and scalable is still one of the main hurdles in allowing widespread usage of decentralised networks. In my opinion this needs to be solved first before blockchains can get to the point where they can create and add large scale value. So I invite you to read the next section to grasp the underlying fundamentals. Because after all these premises need to be true otherwise there isn’t a fundamental case to be bullish on Zilliqa, right?
 
Down the rabbit hole
 
How have they achieved this? Let’s define the basics first: key players on Zilliqa are the users and the miners. A user is anybody who uses the blockchain to transfer funds or run smart contracts. Miners are the (shard) nodes in the network who run the consensus protocol and get rewarded for their service in Zillings (ZIL). The mining network is divided into several smaller networks called shards, which is also referred to as ‘network sharding’. Miners subsequently are randomly assigned to a shard by another set of miners called DS (Directory Service) nodes. The regular shards process transactions and the outputs of these shards are eventually combined by the DS shard as they reach consensus on the final state. More on how these DS shards reach consensus (via pBFT) will be explained later on.
 
The Zilliqa network produces two types of blocks: DS blocks and Tx blocks. One DS Block consists of 100 Tx Blocks. And as previously mentioned there are two types of nodes concerned with reaching consensus: shard nodes and DS nodes. Becoming a shard node or DS node is being defined by the result of a PoW cycle (Ethash) at the beginning of the DS Block. All candidate mining nodes compete with each other and run the PoW (Proof-of-Work) cycle for 60 seconds and the submissions achieving the highest difficulty will be allowed on the network. And to put it in perspective: the average difficulty for one DS node is ~ 2 Th/s equaling 2.000.000 Mh/s or 55 thousand+ GeForce GTX 1070 / 8 GB GPUs at 35.4 Mh/s. Each DS Block 10 new DS nodes are allowed. And a shard node needs to provide around 8.53 GH/s currently (around 240 GTX 1070s). Dual mining ETH/ETC and ZIL is possible and can be done via mining software such as Phoenix and Claymore. There are pools and if you have large amounts of hashing power (Ethash) available you could mine solo.
 
The PoW cycle of 60 seconds is a peak performance and acts as an entry ticket to the network. The entry ticket is called a sybil resistance mechanism and makes it incredibly hard for adversaries to spawn lots of identities and manipulate the network with these identities. And after every 100 Tx Blocks which corresponds to roughly 1,5 hour this PoW process repeats. In between these 1,5 hour no PoW needs to be done meaning Zilliqa’s energy consumption to keep the network secure is low. For more detailed information on how mining works click here.
Okay, hats off to you. You have made it this far. Before we go any deeper down the rabbit hole we first must understand why Zilliqa goes through all of the above technicalities and understand a bit more what a blockchain on a more fundamental level is. Because the core of Zilliqa’s consensus protocol relies on the usage of pBFT (practical Byzantine Fault Tolerance) we need to know more about state machines and their function. Navigate to Viewblock, a Zilliqa block explorer, and just come back to this article. We will use this site to navigate through a few concepts.
 
We have established that Zilliqa is a public and distributed blockchain. Meaning that everyone with an internet connection can send ZILs, trigger smart contracts etc. and there is no central authority who fully controls the network. Zilliqa and other public and distributed blockchains (like Bitcoin and Ethereum) can also be defined as state machines.
 
Taking the liberty of paraphrasing examples and definitions given by Samuel Brooks’ medium article, he describes the definition of a blockchain (like Zilliqa) as:
“A peer-to-peer, append-only datastore that uses consensus to synchronise cryptographically-secure data”.
 
Next he states that: >“blockchains are fundamentally systems for managing valid state transitions”.* For some more context, I recommend reading the whole medium article to get a better grasp of the definitions and understanding of state machines. Nevertheless, let’s try to simplify and compile it into a single paragraph. Take traffic lights as an example: all its states (red, amber and green) are predefined, all possible outcomes are known and it doesn’t matter if you encounter the traffic light today or tomorrow. It will still behave the same. Managing the states of a traffic light can be done by triggering a sensor on the road or pushing a button resulting in one traffic lights’ state going from green to red (via amber) and another light from red to green.
 
With public blockchains like Zilliqa this isn’t so straightforward and simple. It started with block #1 almost 1,5 years ago and every 45 seconds or so a new block linked to the previous block is being added. Resulting in a chain of blocks with transactions in it that everyone can verify from block #1 to the current #647.000+ block. The state is ever changing and the states it can find itself in are infinite. And while the traffic light might work together in tandem with various other traffic lights, it’s rather insignificant comparing it to a public blockchain. Because Zilliqa consists of 2400 nodes who need to work together to achieve consensus on what the latest valid state is while some of these nodes may have latency or broadcast issues, drop offline or are deliberately trying to attack the network etc.
 
Now go back to the Viewblock page take a look at the amount of transaction, addresses, block and DS height and then hit refresh. Obviously as expected you see new incremented values on one or all parameters. And how did the Zilliqa blockchain manage to transition from a previous valid state to the latest valid state? By using pBFT to reach consensus on the latest valid state.
 
After having obtained the entry ticket, miners execute pBFT to reach consensus on the ever changing state of the blockchain. pBFT requires a series of network communication between nodes, and as such there is no GPU involved (but CPU). Resulting in the total energy consumed to keep the blockchain secure, decentralised and scalable being low.
 
pBFT stands for practical Byzantine Fault Tolerance and is an optimisation on the Byzantine Fault Tolerant algorithm. To quote Blockonomi: “In the context of distributed systems, Byzantine Fault Tolerance is the ability of a distributed computer network to function as desired and correctly reach a sufficient consensus despite malicious components (nodes) of the system failing or propagating incorrect information to other peers.” Zilliqa is such a distributed computer network and depends on the honesty of the nodes (shard and DS) to reach consensus and to continuously update the state with the latest block. If pBFT is a new term for you I can highly recommend the Blockonomi article.
 
The idea of pBFT was introduced in 1999 - one of the authors even won a Turing award for it - and it is well researched and applied in various blockchains and distributed systems nowadays. If you want more advanced information than the Blockonomi link provides click here. And if you’re in between Blockonomi and University of Singapore read the Zilliqa Design Story Part 2 dating from October 2017.
Quoting from the Zilliqa tech whitepaper: “pBFT relies upon a correct leader (which is randomly selected) to begin each phase and proceed when the sufficient majority exists. In case the leader is byzantine it can stall the entire consensus protocol. To address this challenge, pBFT offers a view change protocol to replace the byzantine leader with another one.”
 
pBFT can tolerate ⅓ of the nodes being dishonest (offline counts as Byzantine = dishonest) and the consensus protocol will function without stalling or hiccups. Once there are more than ⅓ of dishonest nodes but no more than ⅔ the network will be stalled and a view change will be triggered to elect a new DS leader. Only when more than ⅔ of the nodes are dishonest (>66%) double spend attacks become possible.
 
If the network stalls no transactions can be processed and one has to wait until a new honest leader has been elected. When the mainnet was just launched and in its early phases, view changes happened regularly. As of today the last stalling of the network - and view change being triggered - was at the end of October 2019.
 
Another benefit of using pBFT for consensus besides low energy is the immediate finality it provides. Once your transaction is included in a block and the block is added to the chain it’s done. Lastly, take a look at this article where three types of finality are being defined: probabilistic, absolute and economic finality. Zilliqa falls under the absolute finality (just like Tendermint for example). Although lengthy already we skipped through some of the inner workings from Zilliqa’s consensus: read the Zilliqa Design Story Part 3 and you will be close to having a complete picture on it. Enough about PoW, sybil resistance mechanism, pBFT etc. Another thing we haven’t looked at yet is the amount of decentralisation.
 
Decentralisation
 
Currently there are four shards, each one of them consisting of 600 nodes. 1 shard with 600 so called DS nodes (Directory Service - they need to achieve a higher difficulty than shard nodes) and 1800 shard nodes of which 250 are shard guards (centralised nodes controlled by the team). The amount of shard guards has been steadily declining from 1200 in January 2019 to 250 as of May 2020. On the Viewblock statistics you can see that many of the nodes are being located in the US but those are only the (CPU parts of the) shard nodes who perform pBFT. There is no data from where the PoW sources are coming. And when the Zilliqa blockchain starts reaching their transaction capacity limit, a network upgrade needs to be executed to lift the current cap of maximum 2400 nodes to allow more nodes and formation of more shards which will allow to network to keep on scaling according to demand.
Besides shard nodes there are also seed nodes. The main role of seed nodes is to serve as direct access points (for end users and clients) to the core Zilliqa network that validates transactions. Seed nodes consolidate transaction requests and forward these to the lookup nodes (another type of nodes) for distribution to the shards in the network. Seed nodes also maintain the entire transaction history and the global state of the blockchain which is needed to provide services such as block explorers. Seed nodes in the Zilliqa network are comparable to Infura on Ethereum.
 
The seed nodes were first only operated by Zilliqa themselves, exchanges and Viewblock. Operators of seed nodes like exchanges had no incentive to open them for the greater public.They were centralised at first. Decentralisation at the seed nodes level has been steadily rolled out since March 2020 ( Zilliqa Improvement Proposal 3 ). Currently the amount of seed nodes is being increased, they are public facing and at the same time PoS is applied to incentivize seed node operators and make it possible for ZIL holders to stake and earn passive yields. Important distinction: seed nodes are not involved with consensus! That is still PoW as entry ticket and pBFT for the actual consensus.
 
5% of the block rewards are being assigned to seed nodes (from the beginning in 2019) and those are being used to pay out ZIL stakers.The 5% block rewards with an annual yield of 10.03% translates to roughly 610 MM ZILs in total that can be staked. Exchanges use the custodial variant of staking and wallets like Moonlet will use the non custodial version (starting in Q3 2020). Staking is being done by sending ZILs to a smart contract created by Zilliqa and audited by Quantstamp.
 
With a high amount of DS & shard nodes and seed nodes becoming more decentralised too, Zilliqa qualifies for the label of decentralised in my opinion.
 
Smart contracts
 
Let me start by saying I’m not a developer and my programming skills are quite limited. So I‘m taking the ELI5 route (maybe 12) but if you are familiar with Javascript, Solidity or specifically OCaml please head straight to Scilla - read the docs to get a good initial grasp of how Zilliqa’s smart contract language Scilla works and if you ask yourself “why another programming language?” check this article. And if you want to play around with some sample contracts in an IDE click here. Faucet can be found here. And more information on architecture, dapp development and API can be found on the Developer Portal.
If you are more into listening and watching: check this recent webinar explaining Zilliqa and Scilla. Link is time stamped so you’ll start right away with a platform introduction, R&D roadmap 2020 and afterwards a proper Scilla introduction.
 
Generalised: programming languages can be divided into being ‘object oriented’ or ‘functional’. Here is an ELI5 given by software development academy: > “all programmes have two basic components, data – what the programme knows – and behaviour – what the programme can do with that data. So object-oriented programming states that combining data and related behaviours in one place, is called “object”, which makes it easier to understand how a particular program works. On the other hand, functional programming argues that data and behaviour are different things and should be separated to ensure their clarity.”
 
Scilla is on the functional side and shares similarities with OCaml: > OCaml is a general purpose programming language with an emphasis on expressiveness and safety. It has an advanced type system that helps catch your mistakes without getting in your way. It's used in environments where a single mistake can cost millions and speed matters, is supported by an active community, and has a rich set of libraries and development tools. For all its power, OCaml is also pretty simple, which is one reason it's often used as a teaching language.
 
Scilla is blockchain agnostic, can be implemented onto other blockchains as well, is recognised by academics and won a so called Distinguished Artifact Award award at the end of last year.
 
One of the reasons why the Zilliqa team decided to create their own programming language focused on preventing smart contract vulnerabilities safety is that adding logic on a blockchain, programming, means that you cannot afford to make mistakes. Otherwise it could cost you. It’s all great and fun blockchains being immutable but updating your code because you found a bug isn’t the same as with a regular web application for example. And with smart contracts it inherently involves cryptocurrencies in some form thus value.
 
Another difference with programming languages on a blockchain is gas. Every transaction you do on a smart contract platform like Zilliqa for Ethereum costs gas. With gas you basically pay for computational costs. Sending a ZIL from address A to address B costs 0.001 ZIL currently. Smart contracts are more complex, often involve various functions and require more gas (if gas is a new concept click here ).
 
So with Scilla, similar to Solidity, you need to make sure that “every function in your smart contract will run as expected without hitting gas limits. An improper resource analysis may lead to situations where funds may get stuck simply because a part of the smart contract code cannot be executed due to gas limits. Such constraints are not present in traditional software systems”. Scilla design story part 1
 
Some examples of smart contract issues you’d want to avoid are: leaking funds, ‘unexpected changes to critical state variables’ (example: someone other than you setting his or her address as the owner of the smart contract after creation) or simply killing a contract.
 
Scilla also allows for formal verification. Wikipedia to the rescue:
In the context of hardware and software systems, formal verification is the act of proving or disproving the correctness of intended algorithms underlying a system with respect to a certain formal specification or property, using formal methods of mathematics.
 
Formal verification can be helpful in proving the correctness of systems such as: cryptographic protocols, combinational circuits, digital circuits with internal memory, and software expressed as source code.
 
Scilla is being developed hand-in-hand with formalization of its semantics and its embedding into the Coq proof assistant — a state-of-the art tool for mechanized proofs about properties of programs.”
 
Simply put, with Scilla and accompanying tooling developers can be mathematically sure and proof that the smart contract they’ve written does what he or she intends it to do.
 
Smart contract on a sharded environment and state sharding
 
There is one more topic I’d like to touch on: smart contract execution in a sharded environment (and what is the effect of state sharding). This is a complex topic. I’m not able to explain it any easier than what is posted here. But I will try to compress the post into something easy to digest.
 
Earlier on we have established that Zilliqa can process transactions in parallel due to network sharding. This is where the linear scalability comes from. We can define simple transactions: a transaction from address A to B (Category 1), a transaction where a user interacts with one smart contract (Category 2) and the most complex ones where triggering a transaction results in multiple smart contracts being involved (Category 3). The shards are able to process transactions on their own without interference of the other shards. With Category 1 transactions that is doable, with Category 2 transactions sometimes if that address is in the same shard as the smart contract but with Category 3 you definitely need communication between the shards. Solving that requires to make a set of communication rules the protocol needs to follow in order to process all transactions in a generalised fashion.
 
And this is where the downsides of state sharding comes in currently. All shards in Zilliqa have access to the complete state. Yes the state size (0.1 GB at the moment) grows and all of the nodes need to store it but it also means that they don’t need to shop around for information available on other shards. Requiring more communication and adding more complexity. Computer science knowledge and/or developer knowledge required links if you want to dig further: Scilla - language grammar Scilla - Foundations for Verifiable Decentralised Computations on a Blockchain Gas Accounting NUS x Zilliqa: Smart contract language workshop
 
Easier to follow links on programming Scilla https://learnscilla.com/home Ivan on Tech
 
Roadmap / Zilliqa 2.0
 
There is no strict defined roadmap but here are topics being worked on. And via the Zilliqa website there is also more information on the projects they are working on.
 
Business & Partnerships  
It’s not only technology in which Zilliqa seems to be excelling as their ecosystem has been expanding and starting to grow rapidly. The project is on a mission to provide OpenFinance (OpFi) to the world and Singapore is the right place to be due to its progressive regulations and futuristic thinking. Singapore has taken a proactive approach towards cryptocurrencies by introducing the Payment Services Act 2019 (PS Act). Among other things, the PS Act will regulate intermediaries dealing with certain cryptocurrencies, with a particular focus on consumer protection and anti-money laundering. It will also provide a stable regulatory licensing and operating framework for cryptocurrency entities, effectively covering all crypto businesses and exchanges based in Singapore. According to PWC 82% of the surveyed executives in Singapore reported blockchain initiatives underway and 13% of them have already brought the initiatives live to the market. There is also an increasing list of organisations that are starting to provide digital payment services. Moreover, Singaporean blockchain developers Building Cities Beyond has recently created an innovation $15 million grant to encourage development on its ecosystem. This all suggest that Singapore tries to position itself as (one of) the leading blockchain hubs in the world.
 
Zilliqa seems to already taking advantage of this and recently helped launch Hg Exchange on their platform, together with financial institutions PhillipCapital, PrimePartners and Fundnel. Hg Exchange, which is now approved by the Monetary Authority of Singapore (MAS), uses smart contracts to represent digital assets. Through Hg Exchange financial institutions worldwide can use Zilliqa's safe-by-design smart contracts to enable the trading of private equities. For example, think of companies such as Grab, AirBnB, SpaceX that are not available for public trading right now. Hg Exchange will allow investors to buy shares of private companies & unicorns and capture their value before an IPO. Anquan, the main company behind Zilliqa, has also recently announced that they became a partner and shareholder in TEN31 Bank, which is a fully regulated bank allowing for tokenization of assets and is aiming to bridge the gap between conventional banking and the blockchain world. If STOs, the tokenization of assets, and equity trading will continue to increase, then Zilliqa’s public blockchain would be the ideal candidate due to its strategic positioning, partnerships, regulatory compliance and the technology that is being built on top of it.
 
What is also very encouraging is their focus on banking the un(der)banked. They are launching a stablecoin basket starting with XSGD. As many of you know, stablecoins are currently mostly used for trading. However, Zilliqa is actively trying to broaden the use case of stablecoins. I recommend everybody to read this text that Amrit Kumar wrote (one of the co-founders). These stablecoins will be integrated in the traditional markets and bridge the gap between the crypto world and the traditional world. This could potentially revolutionize and legitimise the crypto space if retailers and companies will for example start to use stablecoins for payments or remittances, instead of it solely being used for trading.
 
Zilliqa also released their DeFi strategic roadmap (dating November 2019) which seems to be aligning well with their OpFi strategy. A non-custodial DEX is coming to Zilliqa made by Switcheo which allows cross-chain trading (atomic swaps) between ETH, EOS and ZIL based tokens. They also signed a Memorandum of Understanding for a (soon to be announced) USD stablecoin. And as Zilliqa is all about regulations and being compliant, I’m speculating on it to be a regulated USD stablecoin. Furthermore, XSGD is already created and visible on block explorer and XIDR (Indonesian Stablecoin) is also coming soon via StraitsX. Here also an overview of the Tech Stack for Financial Applications from September 2019. Further quoting Amrit Kumar on this:
 
There are two basic building blocks in DeFi/OpFi though: 1) stablecoins as you need a non-volatile currency to get access to this market and 2) a dex to be able to trade all these financial assets. The rest are build on top of these blocks.
 
So far, together with our partners and community, we have worked on developing these building blocks with XSGD as a stablecoin. We are working on bringing a USD-backed stablecoin as well. We will soon have a decentralised exchange developed by Switcheo. And with HGX going live, we are also venturing into the tokenization space. More to come in the future.”*
 
Additionally, they also have this ZILHive initiative that injects capital into projects. There have been already 6 waves of various teams working on infrastructure, innovation and research, and they are not from ASEAN or Singapore only but global: see Grantees breakdown by country. Over 60 project teams from over 20 countries have contributed to Zilliqa's ecosystem. This includes individuals and teams developing wallets, explorers, developer toolkits, smart contract testing frameworks, dapps, etc. As some of you may know, Unstoppable Domains (UD) blew up when they launched on Zilliqa. UD aims to replace cryptocurrency addresses with a human readable name and allows for uncensorable websites. Zilliqa will probably be the only one able to handle all these transactions onchain due to ability to scale and its resulting low fees which is why the UD team launched this on Zilliqa in the first place. Furthermore, Zilliqa also has a strong emphasis on security, compliance, and privacy, which is why they partnered with companies like Elliptic, ChainSecurity (part of PwC Switzerland), and Incognito. Their sister company Aqilliz (Zilliqa spelled backwards) focuses on revolutionizing the digital advertising space and is doing interesting things like using Zilliqa to track outdoor digital ads with companies like Foodpanda.
 
Zilliqa is listed on nearly all major exchanges, having several different fiat-gateways and recently have been added to Binance’s margin trading and futures trading with really good volume. They also have a very impressive team with good credentials and experience. They dont just have “tech people”. They have a mix of tech people, business people, marketeers, scientists, and more. Naturally, it's good to have a mix of people with different skill sets if you work in the crypto space.
 
Marketing & Community
 
Zilliqa has a very strong community. If you just follow their Twitter their engagement is much higher for a coin that has approximately 80k followers. They also have been ‘coin of the day’ by LunarCrush many times. LunarCrush tracks real-time cryptocurrency value and social data. According to their data it seems Zilliqa has a more fundamental and deeper understanding of marketing and community engagement than almost all other coins. While almost all coins have been a bit frozen in the last months, Zilliqa seems to be on its own bull run. It was somewhere in the 100s a few months ago and is currently ranked #46 on CoinGecko. Their official Telegram also has over 20k people and is very active, and their community channel which is over 7k now is more active and larger than many other official channels. Their local communities) also seem to be growing.
 
Moreover, their community started ‘Zillacracy’ together with the Zilliqa core team ( see www.zillacracy.com ). It’s a community run initiative where people from all over the world are now helping with marketing and development on Zilliqa. Since its launch in February 2020 they have been doing a lot and will also run their own non custodial seed node for staking. This seed node will also allow them to start generating revenue for them to become a self sustaining entity that could potentially scale up to become a decentralized company working in parallel with the Zilliqa core team. Comparing it to all the other smart contract platforms (e.g. Cardano, EOS, Tezos etc.) they don't seem to have started a similar initiatives (correct me if I’m wrong though). This suggest in my opinion that these other smart contract platforms do not fully understand how to utilize the ‘power of the community’. This is something you cannot ‘buy with money’ and gives many projects in the space a disadvantage.
 
Zilliqa also released two social products called SocialPay and Zeeves. SocialPay allows users to earn ZILs while tweeting with a specific hashtag. They have recently used it in partnership with the Singapore Red Cross for a marketing campaign after their initial pilot program. It seems like a very valuable social product with a good use case. I can see a lot of traditional companies entering the space through this product, which they seem to suggest will happen. Tokenizing hashtags with smart contracts to get network effect is a very smart and innovative idea.
 
Regarding Zeeves, this is a tipping bot for Telegram. They already have 1000s of signups and they plan to keep upgrading it for more and more people to use it (e.g. they recently have added a quiz features). They also use it during AMAs to reward people in real time. It’s a very smart approach to grow their communities and get familiar with ZIL. I can see this becoming very big on Telegram. This tool suggests, again, that the Zilliqa team has a deeper understanding what the crypto space and community needs and is good at finding the right innovative tools to grow and scale.
 
To be honest, I haven’t covered everything (i’m also reaching the character limited haha). So many updates happening lately that it's hard to keep up, such as the International Monetary Fund mentioning Zilliqa in their report, custodial and non-custodial Staking, Binance Margin, Futures & Widget, entering the Indian market, and more. The Head of Marketing Colin Miles has also released this as an overview of what is coming next. And last but not least, Vitalik Buterin has been mentioning Zilliqa lately acknowledging Zilliqa and mentioning that both projects have a lot of room to grow. There is much more info of course and a good part of it has been served to you on a silver platter. I invite you to continue researching by yourself :-) And if you have any comments or questions please post here!
submitted by haveyouheardaboutit to CryptoCurrency [link] [comments]

Full overview of Eth 2.0 & 1.x roadmaps from Messari

Full section on Messari's Ethereum trends for 2020 here

ETH 2.0 Research/Governance/Roadmap at a glance

If history is any guide, we’re not going to see ETH 2.0 until 2022 at the earliest, even if the earliest phases of “Serenity” begin getting pushed in mid-2020. ETH 2.0’s rollout breaks down into seven (7!!!) phases and brings with it the promise of staking, sharding, a new virtual machine, and more dancing badgers.
(One of our analysts, Wilson Withiam, put together an excellent overview of both the ETH 2.0 and ETH 1.x roadmaps for this report. They are critical to track and understand at a high-level given how much Ethereum’s performance will affect other competitive projects and most of the DeFi and Web 3 infrastructure. So these next two sections are longer and more technical.)
Here’s what you need to know about the current game plan for crypto’s largest platform.
Phase 0 marks the launch of the “beacon chain”, which will serve as the backbone for a new blockchain. The beacon chain will manage network validators (large early stakers like ConsenSys) and ultimately assign validators to individual shards (slicing the new blockchain into smaller chunks is a key, difficult, controversial scaling decision that’s been made). The new chain will support Ethereum’s new proof-of-stake consensus mechanism, and offer inflation rewards with new ETH2 for those that pony up and lock 32 ETH1 tokens into an irreversible contract. That one way bridge into the new system is also contentious, but it means ETH1 supply will start getting “effectively burned” once token holder begin claiming beacon chain validator slots. Initial reports claimed Jan. 3 as a realistic launch date (lol). It will be amazing to see this launched by end of June.
Phase 1 will introduce 64 individual shard chains (reduced from 1,024!!!) to the network, with the option to increase the total down the road as the design gets tested. The Ethereum elite see sharding as the “key to future scalability” as shards can parallelize transaction processing, something that could improve network performance and reduce individual validator’s costs (good for decentralization). It comes with big risk: this is still theoretical. No network the size of Ethereum has successfully sharded its blockchain. In Phase 1, shard chains will only contain simple data sets (no smart contracts or transaction executions) to test the system’s structure. As with Phase 0, the beacon chain will continue to run in parallel with ETH 1.x throughout the phase. Don’t expect Phase 1 anytime before 2021.
Phase 2 marks the full launch of the ETH2 chain, allowing for on-chain contract execution and introducing the new eWASM virtual machine (dubbed EVM 2.0). At this point, existing dApps can start migrating their contracts from ETH 1.x to a specific shard (one shard per contract) in the new network. Storage rent, charging contract owners for storing data on the network (more on this below), is in the cards as well, which would require mass contract rewrites. Even though Phase 2 intends to replace the original Ethereum blockchain entirely, ETH 1.x may still live on as a shard within ETH2. (How confused are you by now? See why bitcoin will still dominate the macro narrative for a while?) A late 2021 release for Phase 2 is optimistic. Before the end of 2022 would be a win.
The final four phases are less defined, and without an attached timeline:
Phase 3 implements state-minimized clients (because stateless clients are just too much). Phase 4 allows for cross-shard transactions. Phase 5 improves network security and the availability of data proofs. Phase 6 introduces meta-shards, as in “shards within shards within shards,” for near-infinite scaling. If you’re scratching your head and are sadistic enough to read more, the Sharding Wiki page does note, “this may be difficult.”
Scaling and compilation efficiencies aside, the most notable change in Ethereum’s metamorphosis is the transition from proof-of-work to proof-of-stake. PoW is the more battle tested security model for blockchain networks, while PoS may prove to be more efficient but with new and less obvious attack vectors. For the more technical, we recommend reading Bison Trails’ Viktor Bunin on the subject of PoS security threats.
Past research has also shown PoS requires an extra layer of “trust” vs. PoW, to help nodes sync to the network. Most models share specific characteristics to address this trust issue, such as allowing for a dynamic set of validators (rotate your security), promoting token holder participation in consensus, and assessing steep penalties (slashing) for any network participant that violates the protocol guidelines. ETH 2.0 will function similarly, but may be able to learn from other PoS networks (and their R&D) as well as those come live and see real world issues. As Vitalik points out, recent research in PoS resulted in “great theoretical progress,” But...
Listen, we're talking about practice. Not a game. Not a game. Not a game. We're talking about practice. Not a game….Practice? We're talking about practice, man? We're talking about practice. We're talking about practice. We ain't talking about the game. We're talking about practice, man.
Vitalik was eight when this happened, so the clip might help and prove metaphoric.

2 ETH 1.x Research/Governance/Roadmap at a glance.

Ok, one more. Bear with us. Let’s reiterate, ETH 2.0 is a brand new blockchain. It’s going to be a chaotic and high-risk transition. In the meantime, the existing network needs to run existing applications (particularly financial settlements for DeFi transactions). More critical upgrades are needed in the current system.
To that end, ETH 1.x devs have three goals to boost performance and reduce blockchain bloat: (1) introduce client optimizations that increase transaction capacity; (2) cap disk space requirements and prune old, memory-sucking data (so running a node is less expensive and more decentralized); and (3) upgrade the EVM to eWASM, a newer open standard for code compilers that simplifies debugging, and is also used by all the newer smart contract platforms. ETH 1.x developers have decided to split the major tasks amongst four working groups:

Core developers intend to introduce most of these implementations through a series of hard forks, the latest of which activated just over a week ago (Istanbul, Dec. 7). However, Istanbul’s second phase, tentatively scheduled for Q2 next year, has Ethereans at each other’s throats. The controversy boils down to the fork’s inclusion of ProgPoW, an ASIC-resistant hashing algorithm designed to replace Ethereum’s current algo. ProgPoW aims to even the playing field for GPU miners and ward off the entrance of potential ASIC competitors. The miners like that. But many miners and investors see ProgPoW as a threat to their investments. For miners, the change would shift the power dynamic away from mining farms and render expensive, specialized mining hardware useless. Ethereum (and ERC-20) investors intent on securing their assets might balk because ASIC miners typically prop up hash rates (overall chain security) and their costs “naturally create a price-floor for ASK prices of miners’ sell-orders.”
This saga is far from over. The infighting will likely continue leading up to ProgPoW’s activation date mid-next year, and presents the strongest potential for a network split since “The DAO” fork that spawned Ethereum Classic. The looming transition to ETH 2.0 (and proof-of-stake) will likely deter investor pushback, because it’s a short-term battle in a war the miners are ultimately going to lose, anyway.
Unless the roadmap changes back to supporting a hybrid PoW/PoS system, of course, but... Oh my god, I’m just kidding. This section is mercifully over.
submitted by CryptigoVespucci to ethereum [link] [comments]

Monero, the Most Private Cryptocurrency

Monero, the Most Private Cryptocurrency
Written by the CoinEx Institution, this series of jocular and easy to understand articles will show you everything you need to know about major cryptocurrencies, making you fully prepared before jumping into crypto!

https://preview.redd.it/ryvcznqspe451.jpg?width=720&format=pjpg&auto=webp&s=5fa91e26288d7b0a624113ed21172cc9fd5624a3
Monero, or XMR for short, is an open-source cryptocurrency that is safe, reliable, private, and untraceable. It can run on Windows, Mac, Linux, and FreeBSD, and is known as one of the most private cryptocurrencies. In 2018, Monero already ranked 10th in terms of trading volume, with its market value beyond 1 billion US dollars, an evidence for its great fame in this field.
By a special method in cryptography, Monero ensures that all transactions remain 100% irrelevant and untraceable. Perhaps after reading this article, you will understand why it is so special and popular in the increasingly transparent and traceable cryptocurrency circle (After all privacy comes first!).
In fact, many large cryptocurrencies in the world are not anonymous. All transactions on Bitcoin and Ethereum are made public and traceable, which means that anyone can eavesdrop on transactions flowing into and out of the wallet. That has given rise to a new type of cryptocurrency called “privacy currency”! These “privacy currencies” hide encrypted transactions by adopting specific types of passwords. One typical example is Monero, one of the largest privacy cryptocurrencies in the world.
Monero was created on April 18, 2014 under the name BitMonero, literally the combination of Bit (Bitcoin) and Monero (the “coin” in Esperanto). In five days, the community decided to change its name to Monero.
Interestingly, Monero’s creators valued personal privacy and tried to behave in a low-key manner with pseudonyms instead of the real names. It is said that the Monero major contributor’s nickname is “thankful for today”, yet this guy has gradually disappeared from public view as Monero developed day by day.
Unlike many cryptocurrencies derived from BTC, Monero is based on the CryptoNote protocol. It is also the first branch based on the Bytecoin of CryptoNote currency. Here is some information about Bytecoin: BCN, for short, is a decentralized cryptocurrency with a high degree of privacy; it has open-source codes that allow everyone to contribute to the development of the Bytecoin network; and the Bytecoin network provides global users with instant private transactions that are not traceable and at no additional cost.
Yet, as a branch of BCN, Monero outshines its parent in reputation by being different in two ways. First, Monero’s target block time was reduced from 120 seconds to 60 seconds; second, the issuance speed was cut by 50% (which reverted to 120-second residence later, with the issuance time maintained and the reward for each new block doubled). By the way, during the fork, the Monero developers also found a lot of low-quality codes and then refactored them. (That is exactly what geeks will do)
Monero’s modular code structure was also highly appreciated by Wladimir J. van der Laan, one of the core maintainers of Bitcoin.
Monero values privacy, decentralization and scalability, and there are significant algorithm differences in blockchain fuzzification, which sets it apart from its peers. How private is it? Here are more details.
1. Safe and reliable
For a decentralized cryptocurrency, decentralization means that its network is operated by users; transactions are confirmed by decentralized consensus and then recorded on the blockchain irrevocably. Monero needs no third party to guarantee the safety of funds;
2. Privacy protection
Monero confuses all transaction sources, amounts, and recipients through ring signatures, ring confidential transactions, and invisible addresses. Apart from all the advantages of a decentralized cryptocurrency, it is by no means inferior in safeguarding privacy;
3. Unable to track
The sender, the receiver and the transaction amount of all Monero transactions must be anonymous by default. The information on the Monero Blockchain cannot be matched with physical individuals or specific users, so there is no trace to track;
4. Scalable
Everyone knows that Bitcoin’sability to process transactions has always been limited by the scalability issue; as we have mentioned before in the introduction of Bitcoin, the block size of 1MB makes things difficult. But Monero’s developers have created a system that allows the network to process more transactions when needed; what’s more, Monero does not have any “pre-set” restrictions on block size.
Of course, this also means that some malicious miners may block the system with large blocks. To prevent this from happening, Monero has worked out countermeasures: the block reward penalty of the system.
On October 18, 2018, Monero’s latest hard fork changed the consensus mechanism algorithm to CrypotoNight V8. In this hard fork, it introduced the BulletProff bulletproof protocol, which can also effectively reduce the transaction fee of miners without disclosing transactions
It is said that Monero will issue about 18.4 million XMR in around 8 years. Moreover, it eclipses its counterparts in distribution — with no pre-mining or pre-sale, all block rewards will be left to miners by means of the POW mechanism.
Here is the reward scheme of Monero in two stages:
  1. Acceleration: mine 18132000 XMR before May 2022;
  2. Deceleration: Deceleration starts right after 18132000 XMR are mined, and there will be a reward of 0.6XMR for each block mined afterwards. In this way, the overall supply will be kept on a small scale and decelerated.
Monero is also excellent in its development concept that is designed to be anti-ASIC from the very beginning. Here is a brief introduction to ASIC (Special Application Integrated Circuit).
Due to the specificity of ASICs, specially designed ASICs can usually have much higher hashrate than general CPUs, GPUs, and even FPGAs — that makes hashrate excessively centralized and makes it vulnerable to the monopoly of single centralized institutions. Yet the cryptonight algorithm used by Monero allows most CPUs and even FPGAs to get involved and get mining rewards, instead of making GPU the only one that can efficiently mine.
In other words, Monero’s core development team will modify the consensus mechanism algorithm and have a hard fork after some time to ensure its strength against ASIC and the monopoly of hashrate.
However, although Monero has been designed against ASICs to avoid centralization, nearly 43% of its hashrate is still owned by 3 mining pools; in addition, it is not a BTC-based currency, making it even harder to introduce some elements. Of course, Monero is not that newbie-friendly, and thus has not been widely accepted.
Yet each cryptocurrency has its own features. As long as Monero keeps improving its privacy, it will definitely attract increasing followers. If you are interested in Monero, welcome to CoinEx for exchange or trade.

About CoinEx

As a global and professional cryptocurrency exchange service provider, CoinEx was founded in December 2017 with Bitmain-led investment and has obtained a legal license in Estonia. It is a subsidiary brand of the ViaBTC Group, which owns the fifth largest BTC mining pool, which is also the largest of BCH mining, in the world.
CoinEx supports perpetual contract, spot, margin trading and other derivatives trading, and its service reaches global users in nearly 100 countries/regions with various languages available, such as Chinese, English, Korean and Russian.
Website: https://www.coinex.com/
Twitter: https://twitter.com/coinexcom
Telegram: https://t.me/CoinExOfficialENG
Click here to register on CoinEx!
submitted by CoinEx_Institution to Coinex [link] [comments]

Andreas Antonopoulos gets "Satoshi's Vision" completely wrong and shows his misunderstanding of the system. He thinks 1 cpu 1 vote means 1 user 1 vote, a common mistake from people on the Core side.

In this video at the 6m20s mark Andreas Antonopoulos speaks about Satoshi's vision. He speaks about "1 cpu 1 vote" saying that Satoshi designed the system to be decentralized as possible, but Andreas completely misunderstands the meaning of 1 cpu 1 vote. He is falling into the common trap of conflating 1cpu 1 vote with 1 user 1 vote.
Andreas, haven't you even read nChains paper about POW and Theory of the Firm? A cpu is an economic resource:
One of the little-known aspects of bitcoin is the nature of the proof of work system. There are many people, especially those who support a UASF or PoW change that believe a distributed system should be completed as a mesh. In this, they confuse centralised systems with centrality. The truth of the matter, no matter which proof of work system is implemented, they all follow a maximal growth curve that reflects the nature of the firm as detailed in 1937 by Ronald Coase (1937).
The bitcoin White Paper was very specific. users of the system "vote with their CPU power" [1]. What this means, is that the system was never generated to give one vote per person. It is designed purely around economic incentives individuals with more hash power will have provided more investment into the system. These individuals who invest more in the system gain more say in the system. At the same time, no one or even two individuals can gain complete control of the system. We'll explore the nature of cartels in a separately, but these always fail without government intervention. The reason for cartels failing comes down to the simple incentivisation of the most efficient member. The strongest cartel member always ends up propping up the weakest. This leads to a strategy of defection.
No proof of work-based solution ever allows for a scenario where you have one vote to one person. The anti-sybiling functions of bitcoin and all other related systems based on proof of work or similar derivatives are derived from an investment based strategy. Solutions to the implementation of ASIC based systems are constantly proposed as a methodology of limiting the centralisation of proof of work systems as it is termed. The truth of the matter is that the mining function within any proof of work system naturally aligns to business interests. This leads to corporations running machines within data centres. On the way that democracies and republics have migrated away from small groups of people individually voting for an outcome towards a vote for a party, the transactional costs associated with individual choice naturally leads to corporate solutions. In this, the corporation mirrors a political party.
In this paper, we address the issues of using alternate approval work systems with regards to either incorporating alternate functions in an extension of simply securing the network against the use of proof of work systems to create a one person one vote scenario in place of economic incentivisation. We will demonstrate conclusively that all systems migrate to a state of economic efficiency. The consequence of this is that systems form into groups designed to maximise returns. The effect is that bitcoin is not only incentive compatible but is optimal. No system can efficiently collapse into an order of one vote one individual and remain secure. In the firm-based nature of bitcoin, we demonstrate that the inherent nature of the firm is reflected within mining pools. Multiple aggregation strategies exist. The strategies range from the creation of collective firms where members can easily join or leave (mining pools) through to more standard corporate structures
Proof of Work as it relates to the theory of the firm. that are successful within any proof of work system. The system was determined to be based on one- vote per CPU (Satoshi, 2008) and not one vote per person or one vote per IP address. The reasons for this is simple, there is no methodology available that can solve byzantine consensus on an individual basis. The solution developed within bitcoin solves this economically using investment. The parties signal their intent to remain bound to the protocol through a significant investment. Those parties that follow the protocol are rewarded. The alternative strategy takes us back to the former and failed systems such as e-cash that could not adequately solve Sybil attacks and decentralise the network. Bitcoin manages to maintain the decentralise nature of the network through a requirement that no individual party can ever achieve more than 50% of the network hash rate.
In all proof of work systems, there are requirements to inject a costly signal into the network that is designed as the security control. To many people, they believe that the cryptographic element, namely the hashing process is the security feature of bitcoin. This is a fallacy, it is the economic cost that is relevant to the overall system and not the individual element.
The benefits of a hash function are that they are difficult to solve in the nature of the proof of work algorithm but are easy to verify. This economic asymmetry is one of the key features of bitcoin. Once a user has found a solution, they know it can be quickly broadcast and verified by others. Additionally, the hash algorithm provides a fair distribution system based on the amount of invested hash rate. The distinction from proof of stake solution as has been proposed comes in the requirement to constantly reinvest. A proof of stake system requires a single investment. Once this investment is created, the system is incentivised towards the protection of the earlier investment. This leads to a scenario known as a strategic oligopoly game.
The solution using a proof of work algorithm is the introduction of an ongoing investment. This is different to an oligopoly game in that sunk cost cannot make up for continued investment. In a proof of stake system, prior investment is crystallised allowing continued control with little further investment. Proof of work differs in that it requires continuous investment. More than this, it requires innovation. As with all capitalist systems, they are subject to Schumpeterian dynamical change (Shumpeter, 1994). The system of creative destruction allows for cycles of innovation. Each innovation leads to waves of creation over the destruction of the old order.
This process creates continued growth. Proof of work-based systems continue to grow and continue to update and change. Any incumbent corporation or other entity needs to continue to invest knowing that their continued dominance is not assured. In bitcoin, we have seen innovative leaps as people moved from CPU-based mining into GPU-based systems. This initial innovation altered the software structure associated with the mining process in bitcoin. That change significantly altered the playing field leading to novel techniques associated with FPGAs and later ASICs dedicated to a specific part of the mining process.
The error held by many people is that this move from a CPU-based solution into more costly implementations could have been averted. A consequence of this has been the introduction of alternative proof of work systems into many of the alt-coins
These systems have been implemented without the understanding that it is not the use of ASICs that is an issue. It is that the belief that individual users can individually mine in a mesh system will be able to be implemented as a successful proof of work. In the unlikely event that a specialised algorithm was implemented that could only run once on any one machine CPU, it would still lead to the eventual creation of corporate data centres for mining. In the section above, we showed using Arrow’s theorem how only a single use proof of work system can be effective. If we extend this and look at the Theory of the Firm (Coase, 1937) we note that in a system in Litecoin and Dogecoin for example. A00137:
Proof of Work as it relates to the theory of the firm. of prices, reduction could be carried out without any organisation. One issue against this arises from the cost of information. Interestingly, as we move into a world of increasingly more information, it becomes scarce information that is important. As the amount of information becomes more voluminous, the ability to uncover accurate and timely information becomes scarcer. The ability to specialise in the coordination of the various factors of production and the distribution of information leads towards vertical integration within firms. We see this first voiced in Adam Smith’s (Smith, 1776) postulation on the firm:
Everyone can choose to either seek further information or act on the information that they already have. This information can be in the form of market knowledge, product knowledge, or expertise, but at some point, the individual needs to decide to act. There is a cost to obtaining information. The returns on obtaining more information hit a maximum level and start to decrease at a certain point. The entrepreneur acts as a guiding influence managing the risk associated with incomplete information compared to the risk of not acting but rather waiting to obtain more information.
In the instance of bitcoin mining, the firm can increase in size through the integration of multiple specialist roles. Even given the assumption that any one process can run on but a single CPU, we come to the scenario of high-end datacentre servers. The Intel Xeon Phi 7290f implements 72 Atom CPU Cores. Each core runs two threads. Even taking the control system into account, this leaves 142 processes able to run per system. With four cards per RU this allows for datacentre implementations of 5,964 mining processes to run on a pure CPU-based proof of work implementation. One person can manage a small number of mining server implementations within a home or small business environment. In large data centre-based organisations such as Facebook, a single administrator can run 20,000 servers
The effect of this would be one individual managing 2,840,000 individual CPU-based mining processes. This alone is outside the scaling capabilities of any individual. This can be further enhanced as cost savings through the creation of large data centres, management savings and integrating multiple network and systems administrators is considered. As we start to add additional layers we come to a maximum where it is no longer profitable to grow the firm in size. Right up until that point, the firm will grow.
submitted by cryptorebel to btc [link] [comments]

Building a mining rig for 100k USD.

Ok so basically my question is this:
How do you build the most efficient Bitcoin mining rig with 100k USD?
I've been trying to do some research but I really want some info from people who have/have had mining rigs in the past. I wanna know what are some potential issues that could arise?
I want to know if it's most cost effective to buy a bunch of GPU's and set them up with some mining motherboards or is it better to buy a pre-built machine like the AntMiners?
If I'm building a rig for 100k and plan to pay for electricity using the bitcoin mined each month then how much money do I need to have as backup to make sure I can pay for electricity if BTC price drops significantly?
Just anything useful that someone who plans to build a rig should account for.
Also: Yes I acknowledge that BTC is very unstable and that even though mining right now (for 10K/BTC) is profitable it could be the case that later on I'm mining in the negative for several months or years.
Thanks.
submitted by AnimeRegime6987 to CryptoCurrency [link] [comments]

A comprehensive review of miner arguments against issuance reduction

First I wanted to start by saying Hudson Jameson did a phenomenal job wrangling all these different stakeholders to the core devs meeting today and playing the part of an effective, neutral moderator. It was a really interesting meeting and great to hear all viewpoints. I'm sure many of you live streamed it as well.
I’m an Ethereum investor and active user, and I took notes on the most prominent miner arguments against issuance reduction along with my thoughts on each. Would love to hear any thoughts or any ones I may have missed.
GPUs that leave network after issuance reduction can be used to attack Network Security (Xin Xu)
Xin Xu argues that a decline in issuance from 3 to 2 (33%) will cause a drop in hashrate by 33%, and that such a large drop in hashrate will lead to an influx of GPUs on the market that can be used to attack Ethereum. This argument is predicated on the idea that hashrate will drop significantly. However, any drop in hashrate will decrease difficulty so mathematically a 33% drop in issuance should have at most a ~18% impact to total hashrate assuming a linear relationship. I don't believe that a drop in Ethereum Network Hashrate from current levels (280 TH/s) to January 2018 levels (230 TH/s) is a doomsday scenario. And the real drop will certainly be much smaller for two reasons. 1) Historical data shows that hashrate is extremely resilient against drops in price as well as issuance (source: https://pbs.twimg.com/media/DlTEyKBV4AERGtB.jpg:large). 2) Historical data also shows that all Ethereum and Bitcoin issuance reductions were followed by price increases which could partially or completely offset the decline in hashrate.
Issuance Reduction will drive a dramatic shift in hardware composition of the network (Brian Venturo)
Brian Venturo argues that a reduction in issuance will price out GPU miners and cause the network to dramatically shift towards ASIC miners in the short term, increasing mining centralization. However, miners on the call pointed out that currently available ASICS (Antminer E3) is in-line with top GPUs in terms of mining efficiency. It’s only when we compare claims from as-yet unreleased ASIC manufacturers (Innosilicon A10) to 2-year old GPU technology (GTX 1080) that we see any risk of an efficiency gap. Second, the total Ethereum network hashrate is 280 TH/s. This is equivalent to 577,000 Innosilicon A10s, which would cost $3.3 Billion (at $5700 each). Any shift of even 10-20% in Ethereum network hardware composition will be slow and steady, and as we heard on the call, miners looking to spend significant capital on new hardware are considering major ROI headwinds from 1) upcoming shift to PoS and 2) possible exploration of new ASIC-resistant algos like ProgPOW. Both of these would brick current generation ASICS while GPUS would retain their resale value. More work needs to be done exploring ASIC-resistant POW algorithms, and there's no reason why issuance-reduction EIPs should be roadblocked in the interim.
EIP 1295 as an alternative (Brian Venturo)
Brian Venturo cites the current rules around Uncle and Nephew rewards as causing weird incentives that miners are exploiting to maximize uncle rate and squeeze higher issuance out of the network. This is a super interesting point, and one that I would love to see explored in more detail (as the downstream implications could be quite complex) in addition to EIP-1234. There’s no reason why 1295 is mutually exclusive with EIP-1234, and positioning it that way is a clever tactic to delay any issuance reduction. Brian himself suggested an issuance reduction in 2019 on top of EIP-1295.
My Final thought
I am in full support of EIP-1234 as a moderate issuance reduction to reduce Ethereum inflation and the amount we are overpaying miners for security. Looking back on it, last year’s 40% reduction from 5 eth/block to 3 eth/block has turned out to be a phenomenally good decision. Since then, hashrates have increased 3x while price has declined 20% (was $330 pre-fork), all while we reduced inflation by 40%. Another modest issuance reduction is a prudent decision that is a natural step in Ethereum’s growth and consistent with the original vision for inflation. In contrast, a difficulty bomb delay without a corresponding issuance reduction should be viewed as an issuance increase.
The quicker we can get this decision behind us, the better. As long as this question looms, investors will lack confidence in Ethereum’s monetary policy, and mining stakeholders will have massive incentive to decrease Ethereum price until Constantinople to increase the chance they can mine at inflated rates through 2019
submitted by AZA214 to ethtrader [link] [comments]

DISC may be a better choice if you didn’t seize the opportunity of Bitcoin

DISC may be a better choice if you didn’t seize the opportunity of Bitcoin
https://preview.redd.it/di9avv4gqx441.png?width=1211&format=png&auto=webp&s=eb92d1eaffc52764bfd4bb0f0c55dead94aabf26
At present, the mainstream consensus mechanism of the encryption world is the Proof of Work (POW) used by Bitcoin (BTC) and the Proof of Stake (POS) used by EOS. In addition, there are more than 30 kinds of consensus mechanisms such as Proof of Capacity (POC) in the encryption world. Among them, POC has caused extensive discussion. Together with POC, there is also the cryptocurrency Diskcoin in the POC ecosystem, a core algorithm based on POC, which aims to establish a cryptocurrency of a greener energy-saving mining system.
Before introducing POC and Diskcoin, you need to understand POW and BTC first.

Proof Of Work
Ten years ago, Satoshi Nakamoto took the POW into the BTC mining, which once brought more than 200 billion US dollars in economic value to human society. BTC can sit on the throne of the encrypted world, and POW is indispensable. The BTC changed the traditional business model and the role of “miners” was born.
With the development, the miners gradually increased and the competition more fierce. There are large miners, mining pools, mines and other organizations. Due to the emergence of these organizations, the limitations of BTC have gradually emerged. BTC is proud of its decentralization and is no longer decentralized.
At the same time, POW was gradually occupied by ASIC. More and more mining machine manufacturers began to participate in mining. The reward of mining was slowly monopolized by giants such as mining machine manufacturers and mining pools, and small miners lost the opportunity to compete on the same stage.
BTC has been labeled as a privileged label, and now, without enough assets, it is almost impossible to participate in BTC.

Proof Of Capacity
The original intention of Satoshi Nakamoto to design BTC is that everyone can participate in mining and participate in the casting of “currency”. Therefore, in the early BTC mining, only one ordinary computer was needed for calculation work. Now, if you want to use a computer to mine BTC, it is simply a dream. The mining is mainly based on GPU and ASIC mining machines. This makes the cost of mining and the difficulty of mining greatly increase. Most of the hashrate of the BTC full net has been controlled by a few mining pools.
Power energy is also consumed in large quantities. BTC currently has a total hashrate of 43.42EH/S. It is mainly based on the popular S9 ant mining machine currently on the market. The average S9 hashrate is 13TH/S. The converted BTC full net hashrate is equivalent to 3.34 million S9 mining machines! In use, the power of the S9 is 1500 watts, and the power consumption per day is 36 watts. The power consumption of one S9 per month is 1080 watts. If multiplied by 3.34 million machines, the power consumption of the BTC for one year up to 43.2 billion degrees! The BTC full net has consumed more than 159 countries a year, accounting for 0.2% of global electricity consumption!
Can the BTC face a new solution?
At this point, POC boarded the stage of the encrypted world and launched a competition with POW.

Principle of Proof Of Capacity
Before the mining starts, the POC will calculate in advance and the hash value calculated by shabal256 will be stored in the hardisk. This process is called Plotting.
The POC uses a hash algorithm called shabal256. Compared to the B256 sha256, the shabal256 algorithm requires a longer calculation and workload. The benefit of Shabal256 is that miners do not need to perform hash calculations during mining, so the verification process will be relatively fast, and shabal256 will also prevent cheating.
From this we can know that the energy consumption of POC mining is extremely small. Because the repetitive and large amount of calculations are required compared to the POW, the POC simply scans the stored data.
In order to make everyone understand the POC hard drive mining, I will give you a simple example. POC mining process is similar to the lottery in the hardisk, and the process of mining is that you only need to wait for the lottery to draw the prize, you will scan the lottery in the hardisk to confirm whether it is winning, nothing more.
Someone here may have doubts, will it be faster to use an SSD hard drive? The clear answer is, no. Because the POC only cares the capacity of the hardisk, the larger the capacity is, the more lottery tickets you store on the hardisk, the greater the probability that you win the prize, and this has nothing to do with what kind of hardisk you use, what brand of hardisk.
The POC has changed the traditional mining method, which not only reduces the energy consumption of mining, but also lower the threshold of mining, so that everyone can participate.

Diskcoin
Diskcoin, referred to as DISC. DISC, like Bitcoin, has a total of 21 million. Based on the consensus of hardisk capacity, DISC lowers the participation threshold and allows more participation in the casting of coins, which is more conducive to the realization of decentralization.
DISC uses an upgraded version of the POC algorithm - CPOC Conditioned Proof Of Capacity.
Miners need to Stake Diskcoin to get the most reward. The ratio of Staking is not fixed or gradually reduced. Instead, an algorithmic mechanism called DES (Dynamic Equilibrium Staking) is used to adjust the Staking ratio based on the difficulty of mining. DISC's Staking economy model makes it impossible for miners to sell tokens without restrictions. This initiative is also to maintain the healthy and stable development of the DISC ecosystem.
Supply and demand are very important in economics. Supply and demand are the most direct factors for the decline in commodity prices, and the most effective market feedback. DISC's economic model ensures that the circulation of its currency is stable, keeping it in balance with the number of markets.
CPoC mining combines miners with the entire ecological interests, and replaces the originally consumed power resources with tokens as new production materials, so that the entire ecosystem of DISC continues to expand independently, forming a virtuous cycle system.
submitted by Diskcoin to DiskcoinOrg [link] [comments]

How Are Mining Profits Now? November 2019 What Do YOU Need to MINE ONE BITCOIN In 2020?! - YouTube How to Mine Cryptocurrency and Bitcoin on Mac/Windows, using CPU or GPU with MinerGate and NiceHash Bitcoin Mining 2019 - Should We Mine Bitcoin? - YouTube How Much Can You Make Mining Bitcoin With 6X 1080 Ti ...

It is slightly expensive as its an effective combination of the two most powerful GPUs. It has a hash rate of 46 MH/s and a power consumption of 500W. On a daily basis, it has a power cost of 1.44 USD and returns per day of 2.23 USD. This is considered to be one of the top graphics cards for any Ethereum mining hardware. It has an extremely advanced multicore processor, and as it has a liquid ... Best mining rigs and mining PCs for Bitcoin, Ethereum and more By Matt Hanson , Brian Turner 14 September 2020 If you still fancy mining cryptocurrencies, these are the best pre-built rigs and PCs This isn't fit for SE in my opinion. There is no "most cost effective" mining setup; it all depends on your resources, costs, and goals. Even if you provide all of those, much of it is up to personal preference and predictions of future value. If you want configuration ideas, the wiki page should serve nicely. – BinaryMage May 3 '12 at 3:59 When it comes to mining with a greater speed, ASIC miners play a huge role and they are more productive and cost-effective than the traditional Bitcoin miners. Bitcoin mining is effective only when there is a net benefit in regard to productivity and low cost of running. They are high on consuming electricity and there are users who often combine rigs and ASIC chips just to bring the costs ... Besides electricity, mining hardware is the second most important factor when it comes to mining ETH. Older Cards to get on cheap and experiment with GTX 1080 Ti. A GTX 1080 Ti has a good hash rate of 36.0MH/s. Its power cost per day is $0.7632, which is the median cost per day of all the miners on our list. This gives a return per day of -$0 ...

[index] [23025] [14604] [38069] [15720] [37787] [43033] [43740] [27769] [50147] [50235]

How Are Mining Profits Now? November 2019

Quadro P600 Review And Bitcoin Mining With Nicehash Minergate And Betterhash NVIDIA QUADRO P600—REALIZING DEMANDING VISUAL COMPUTING WORKFLOW PERFORMANCE. The NVIDIA Quadro P600 combines a 384 ... Latest Video: http://bit.ly/BW10000 1. Buy Bitcoins: http://bit.ly/BWCoinbase 2. Best Crypto Exchange: http://bit.ly/BWBinance 3. ROBINHOOD http://bit.ly/ROB... SUBSCRIBE FOR MORE HOW MUCH - http://shorturl.at/arBHL Nviddia GTX 1080 Ti - https://amzn.to/2Hiw5xp 6X GPU Mining Rig Case - https://bitcoinmerch.com/produc... So is Bitcoin mining worth it in may 2019? is Bitcoin mining in anyway profitable? Or worth it? should purchase a bitcoin mining machine (antminer or ASIC) or build your first mini bitcoin mining ... So if you are thinking of Mining on Mac, without a GPU the MacBook Pro will get you 20 Bytecoins a day, which is today worth only 10 cents, but with minimal electricity usage. With a GPU like the ...

#