Web 1.0 vs 2.0 vs 3.0.
When talking about Web standards, usually the different Web generations are brought up, what do they actually mean?, most of us are used to browsing on the web and using websites, but we do not understand these terms; let’s see some basic definitions.
Web 1.0 sites were important because they were the first generation of the World Wide Web, and they paved the way for the development of subsequent generations of the Internet. This era of the World Wide Web was defined by static HTML pages that were manually created and updated. Some problems with the firsts websites include the fact that sites were static and required manual updates, which made it difficult to keep content fresh. Additionally, because HTML was the only language used to create Web 1.0 sites, they were not very interactive or user-friendly.
Some Web 1.0 examples include Geocities, Tripod, and Angelfire.
Some Web 2.0 examples include Facebook, Twitter, and LinkedIn.
There is no precise definition for this since this generation is in its early stages but it’s generally said that new applications will be more decentralized and reliant on blockchain technologies, thus giving users more control over their data and removing central authorities that can take advantage of data users provide and or regulate information users consume. Some of the issues encountered in this generation of websites have been lack of standards, lack of adoption, and general complexity.
Web 3.0 Examples include IPFS, Smart Contracts and Cryptocurrencies.
We need to care about web standards because they help ensure that the web is accessible to everyone, regardless of their device or browser. Additionally, web standards help make the web more efficient and easier to use. We hope that this post can give you a general idea of the different eras of the modern Web.
Check out this post in:
Image credit: Denny Müller on Unplash.