Based on our record, Materialize should be more popular than Hadoop. It has been mentiond 72 times since March 2021. We are tracking product recommendations and mentions on various public social media platforms and blogs. They can help you identify which product is more popular and what people think of it.
It's hard to write something that is both accessible and well-motivated. The best uses of category theory is when the morphisms are far more exotic than "regular functions". E.g. It would be nice to describe a circuit of live queries (like https://materialize.com/ stuff) with proper caching, joins, etc. Figuring this out is a bit of an open problem. Haskell's standard library's Monad and stuff are watered down to... - Source: Hacker News / 5 months ago
> [...] `https://materialize.com/` to solve their memory issues [...] Disclaimer: I work at Materialize Recently there have been major improvements in Materialize's memory usage as well as using disk to swap out some data. I find it pretty easy to hook up to Postgres/MySQL/Kafka instances: https://materialize.com/blog/materialize-emulator/. - Source: Hacker News / 6 months ago
I agree. So many disparate solutions. The streaming sql primitives are by themselves good enough (e.g. `tumble`, `hop` or `session` windows), but the infrastructural components are always rough in real life use cases. Crossing fingers for solutions like `https://github.com/feldera/feldera` to solve their memory issues, or `https://clickhouse.com/docs/en/materialized-view` to solve reliable streaming consumption.... - Source: Hacker News / 6 months ago
Or the related Materialize stuff https://materialize.com/. - Source: Hacker News / 7 months ago
The original post makes so much more sense in this context! One of the "holy grails" in my mind is making CQRS and dataflow programming as easy to learn and maintain as existing imperative programming languages - and easy to weave into real-time UX. There are so many backend endpoints in the wild that do a bunch of things in a loop, many of which will require I/O or calls to slow external endpoints, transform the... - Source: Hacker News / 7 months ago
Over the years, Indian developers have played increasingly vital roles in many international projects. From contributions to frameworks such as Kubernetes and Apache Hadoop to the emergence of homegrown platforms like OpenStack India, India has steadily carved out a global reputation as a powerhouse of open source talent. - Source: dev.to / 4 days ago
One of the key attributes of Apache License 2.0 is its flexible nature. Permitting use in both proprietary and open source environments, it has become the go-to choice for innovative projects ranging from the Apache HTTP Server to large-scale initiatives like Apache Spark and Hadoop. This flexibility is not solely legal; it is also philosophical. The license is designed to encourage transparency and maintain a... - Source: dev.to / about 2 months ago
Apache Hadoop is more than just software—it’s a full-fledged ecosystem built on the principles of open collaboration and decentralized governance. Born out of a need to process vast amounts of information efficiently, Hadoop uses a distributed file system and the MapReduce programming model to enable scalable, fault-tolerant computing. Central to its success is a diverse ecosystem that includes influential... - Source: dev.to / 2 months ago
Navya: Designed to streamline administrative processes in educational institutions, Navya continues to demonstrate the power of open source in addressing local needs. Additionally, India’s vibrant tech communities are well represented on platforms like GitHub and SourceForge. These platforms host numerous Indian-led projects and serve as collaborative hubs for developers across diverse technology landscapes.... - Source: dev.to / 2 months ago
The rise of big data has seen Java arise as a crucial player in this domain. Tools like Hadoop and Apache Spark are built using Java, enabling businesses to process and analyze massive datasets efficiently. Java’s scalability and performance are critical for big data results that demand high trustability. - Source: dev.to / 5 months ago
RisingWave - RisingWave is a stream processing platform that utilizes SQL to enhance data analysis, offering improved insights on real-time data.
Apache Spark - Apache Spark is an engine for big data processing, with built-in modules for streaming, SQL, machine learning and graph processing.
Apache Flink - Flink is a streaming dataflow engine that provides data distribution, communication, and fault tolerance for distributed computations.
PostgreSQL - PostgreSQL is a powerful, open source object-relational database system.
Apache Kafka - Apache Kafka is an open-source message broker project developed by the Apache Software Foundation written in Scala.
Apache Storm - Apache Storm is a free and open source distributed realtime computation system.