FileCoin might be a bit more popular than Apache Spark. We know about 78 links to it since March 2021 and only 70 links to Apache Spark. We are tracking product recommendations and mentions on various public social media platforms and blogs. They can help you identify which product is more popular and what people think of it.
WeaveChain will be a CosmosSDK based DePIN blockchain and a marketplace to match database developers / dapps with rollup operators. It's basically a Filecoin for database. zkDB/WeaveDB is to WeaveChain as IPFS is to Filecoin. We will introduce 2 unique components to connect with real-world data and web2. - Source: dev.to / 15 days ago
Abstract: This post explores how tokenization is revolutionizing the sustainability of open-source projects. We dive into the background of open-source funding challenges, define key blockchain and tokenization concepts, discuss core features, and present practical use cases with real-world examples. Furthermore, we analyze the challenges and limitations facing decentralized funding models and conclude with a... - Source: dev.to / 15 days ago
Tokenization brings with it the power to transform how value is created and distributed within a community. One striking example is the Basic Attention Token (BAT). BAT leverages blockchain technology to reward users for their attention. This is not just a novel way to capture value, but also a means to protect individual privacy in a digital age that increasingly relies on data collection and targeted... - Source: dev.to / 3 months ago
For example, decentralized data storage projects like Filecoin, Arweave, and Sia posted 50-100% user growth, providing blockchain-powered alternatives to AWS, Google Cloud, and Dropbox for distributed app data security. - Source: dev.to / over 1 year ago
Filecoin, which is based on IPFS, creates a market for unused storage. I think that idea is great but for adoption it needs to be as simple as Dropbox to store files. But visit [filecoin.io](https://filecoin.io/) and the dropbox-like app that you could be willing to try is nowhere to be found. So maybe it is an enterprise solution? That isn't spelled out either. So I am not surprised that this has little trackion... - Source: Hacker News / over 1 year ago
Apache Iceberg defines a table format that separates how data is stored from how data is queried. Any engine that implements the Iceberg integration β Spark, Flink, Trino, DuckDB, Snowflake, RisingWave β can read and/or write Iceberg data directly. - Source: dev.to / about 1 month ago
Apache Spark powers large-scale data analytics and machine learning, but as workloads grow exponentially, traditional static resource allocation leads to 30β50% resource waste due to idle Executors and suboptimal instance selection. - Source: dev.to / about 1 month ago
One of the key attributes of Apache License 2.0 is its flexible nature. Permitting use in both proprietary and open source environments, it has become the go-to choice for innovative projects ranging from the Apache HTTP Server to large-scale initiatives like Apache Spark and Hadoop. This flexibility is not solely legal; it is also philosophical. The license is designed to encourage transparency and maintain a... - Source: dev.to / 2 months ago
[1] S. Russell and P. Norvig, Artificial Intelligence: A Modern Approach. Pearson, 2020. [2] F. Chollet, Deep Learning with Python. Manning Publications, 2018. [3] C. C. Aggarwal, Data Mining: The Textbook. Springer, 2015. [4] J. Dean and S. Ghemawat, "MapReduce: Simplified Data Processing on Large Clusters," Communications of the ACM, vol. 51, no. 1, pp. 107-113, 2008. [5] Apache Software Foundation, "Apache... - Source: dev.to / 2 months ago
If you're designing an event-based pipeline, you can use a data streaming tool like Kafka to process data as it's collected by the pipeline. For a setup that already has data stored, you can use tools like Apache Spark to batch process and clean it before moving ahead with the pipeline. - Source: dev.to / 3 months ago
IPFS - IPFS is the permanent web. A new peer-to-peer hypermedia protocol.
Apache Flink - Flink is a streaming dataflow engine that provides data distribution, communication, and fault tolerance for distributed computations.
Sia - Sia is a decentralized cloud object storage where mutually-distrusting parties interact directly creating a trustless cloud storage marketplace without intermediaries, borders, vendor lock-ins, spying, throttling or walled gardens.
Hadoop - Open-source software for reliable, scalable, distributed computing
Storj Object Storage - Storj Distributed Cloud Object Storage Global is an object storage which is fully compatible with Amazon S3, globally distributed in nature, automatically decentralized, always encrypted and lightning fast through parallelization.
Apache Storm - Apache Storm is a free and open source distributed realtime computation system.