Based on our record, Minio seems to be a lot more popular than Hadoop. While we know about 154 links to Minio, we've tracked only 15 mentions of Hadoop. We are tracking product recommendations and mentions on various public social media platforms and blogs. They can help you identify which product is more popular and what people think of it.
Here are the basic steps to getting a minio tenant deployed inot kubernetes. There are some pre-requisites tasks to be deployed (and will not be covered in this article) including. - Source: dev.to / 26 days ago
I'd throw minio [1] in the list there as well for homelab k8s object storage. [1] https://min.io/. - Source: Hacker News / 4 months ago
Can you just append the data to a blob using something like the s3 blob api? AWS, Azure and Minio https://min.io/ all support it. That way you don't have to reinvent the wheel. Source: 8 months ago
With that being said, you better take a look at something more WAN optimized and more secure, like S3 storage. You can build the S3 storage (and gain immutability) using something like MinIO (https://min.io/) or Ceph (https://ceph.io/en/) or check out Object First Ootbi offerings - https://objectfirst.com/object-storage/ (I work for them). Source: 9 months ago
ToolJet allows you to build applications that use relational and non-relational databases, REST APIs, OpenAI technologies, and cloud storage like Google Cloud Storage, AWS S3, and Minio. It is an excellent development tool helping individuals, developers, and businesses create and ship products faster. - Source: dev.to / 9 months ago
Did you check out tools like https://hadoop.apache.org/ ? Source: about 1 year ago
There are different ways to implement parallel dataflows, such as using parallel data processing frameworks like Apache Hadoop, Apache Spark, and Apache Flink, or using cloud-based services like Amazon EMR and Google Cloud Dataflow. It is also possible to use parallel dataflow frameworks to handle big data and distributed computing, like Apache Nifi and Apache Kafka. Source: about 1 year ago
There are several frameworks available for batch processing, such as Hadoop, Apache Storm, and DataTorrent RTS. - Source: dev.to / over 1 year ago
A copy of Hadoop installed on each of these machines. You can download Hadoop from the Apache website, or you can use a distribution like Cloudera or Hortonworks. - Source: dev.to / over 1 year ago
The Apache™ Hadoop™ project develops open-source software for reliable, scalable, distributed computing. - Source: dev.to / over 1 year ago
Ceph - Ceph is a distributed object store and file system designed to provide excellent performance...
Apache Spark - Apache Spark is an engine for big data processing, with built-in modules for streaming, SQL, machine learning and graph processing.
Google Cloud Storage - Google Cloud Storage offers developers and IT organizations durable and highly available object storage.
Apache Cassandra - The Apache Cassandra database is the right choice when you need scalability and high availability without compromising performance.
Azure Blob Storage - Use Azure Blob Storage to store all kinds of files. Azure hot, cool, and archive storage is reliable cloud object storage for unstructured data
Apache Flink - Flink is a streaming dataflow engine that provides data distribution, communication, and fault tolerance for distributed computations.