What is Kappa Architecture?

Kappa Architecture – Where Every Thing Is A Stream (pathirage.org)

Kappa Architecture is a software architecture pattern. Rather than using a relational DB like SQL or a key-value store like Cassandra, the canonical data store in a Kappa Architecture system is an append-only immutable log. From the log, data is streamed through a computational system and fed into auxiliary stores for serving.

Kappa Architecture is a simplification of Lambda Architecture. A Kappa Architecture system is like a Lambda Architecture system with the batch processing system removed. To replace batch processing, data is simply fed through the streaming system quickly.

But why?

Kappa Architecture revolutionizes database migrations and reorganizations: just delete your serving layer database and populate a new copy from the canonical store! Since there is no batch processing layer, only one set of code needs to be maintained.

Says who?

The idea of Kappa Architecture was first described in an article by Jay Kreps from LinkedIn. Then came the talk “Turning the database inside out with Apache Samza” by Martin Kleppmann at 2014 StrangeLoop which inspired this web site.

TURNING THE DATABASE INSIDE OUT WITH APACHE SAMZA

HOW DO I MAKE MY OWN?

RESOURCES

Tools

LOG DATA STORES

An append-only immutable log store is the canonical store in a Kappa Architecture (or Lambda Architecture) system. Some log databases:

STREAMING COMPUTATION SYSTEMS

In Kappa Architecture, data is fed from the log store into a streaming computation system. Some distributed streaming systems:

SERVING LAYER STORES

The purpose of the serving layer is to provide optimized responses to queries. These databases aren’t used as canonical stores: at any point, you can wipe them and regenerate them from the canonical data store. Almost any database, in-memory or persistent, might be used in the serving layer. This also includes special-purpose databases, e.g. for full text search.

Strange Loop

Strange Loop is a multi-disciplinary conference that brings together the developers and thinkers building tomorrow’s technology in fields such as emerging languages, alternative databases, concurrency, distributed systems, security, and the web.

Strange Loop was created in 2009 by software developer Alex Miller and is now run by a team of St. Louis-based friends and developers under Strange Loop LLC, a for-profit venture.

Some of our guiding principles:
No marketing. Keynotes are never sold to sponsors. The conference mailing lists are never sold or given to sponsors.
Tech, not process. Talks are in general code-heavy, not process-oriented (agile, testing, etc). There are many fine speakers, topics, and conferences in the process area. This is not one of them.
Technology stew. Interesting stuff happens when you get people from different areas in the same room. Strange Loop has a broad range of topics from academia, industry, and a touch of weirdness.

About – Strange Loop (thestrangeloop.com)

Apache Impala vs Presto

Apache Impala vs Presto: What are the differences?

What is Apache Impala? Real-time Query for Hadoop. Impala is a modern, open source, MPP SQL query engine for Apache Hadoop. Impala is shipped by Cloudera, MapR, and Amazon. With Impala, you can query data, whether stored in HDFS or Apache HBase – including SELECT, JOIN, and aggregate functions – in real time.

What is Presto? Distributed SQL Query Engine for Big Data. Presto is an open source distributed SQL query engine for running interactive analytic queries against data sources of all sizes ranging from gigabytes to petabytes.

Apache Impala and Presto belong to „Big Data Tools“ category of the tech stack.
https://stackshare.io/stackups/impala-vs-presto

stackshare.io

Some links about the good old (and dead?) SSAS Multidimensional Cube

Why doesn’t SSAS cache the entire cube?
Also, if you cube is much larger than the memory available to SSAS, then you would expect to see continual IO, and it is likely to be quite well optimised. However, when you have a 64 bit server with a cube that is larger than 3GB but is comfortably less than the server memory, you might be surprised to see the volume of continual IO.

http://richardlees.blogspot.com/2011/12/why-doesnt-ssas-cache-entire-cube.html

Best Practices for Performance Tuning in SSAS Cubes, you are in right place. Define cascading attribute relationships, for example, day > Month > Quarter > year  and define user hierarchies of related attributes (called natural hierarchies) within each dimension as Appropriate for your data

Best Practices for Performance Tuning in SSAS Cubes, you are in right place. Define cascading attribute relationships, for example, day > Month > Quarter > year  and define user hierarchies of related attributes (called natural hierarchies) within each dimension as Appropriate for your data

Remove redundant relationships between attributes to assist the query execution engine in generating the appropriate query plan. Attributes need to have either a direct or an indirect relationship to key attributes, not both.


https://mindmajix.com/msbi/best-practices-for-performance-tuning-in-ssas-cube

https://kejserbi.wordpress.com/

https://hub.packtpub.com/query-performance-tuning-microsoft-analysis-services-part-1/

https://github.com/RichieBzzzt/SSASActivityMonitor

https://github.com/ssasdiag/SSASDiag

https://christianb7.wordpress.com/2012/11/11/analysis-services-2012-configuration-settings/

https://www.mssqltips.com/sqlservertip/2568/ssas-best-practices-and-performance-optimization-part-4-of-4/

Aardvark’d: 12 Weeks with Geeks

Aardvark’d: 12 Weeks with Geeks is a 2005 documentary film about the development of Fog Creek Copilot, a remote assistance software tool. Conceptualization of the film began when Fog Creek Software CEO Joel Spolsky announced on his blog that he was seeking a filmmaker to document the development of the product, then called Project Aardvark.[1][2]

https://en.wikipedia.org/wiki/Aardvark%27d:_12_Weeks_with_Geeks

https://www.imdb.com/title/tt0813987/