Giter VIP home page Giter VIP logo

gridql's Introduction

CI codecov

gridql

Scaffolding for converting event documents into a mesh

GraphQL has a Conway's Law problem: its built by a company that has lots of smaller business units that are controlled, synchronised and presetend by a single umbrella company Meta. So they solved the distributed nature of GraphQL with a single point of failure Apollo.

Apollo Server has lots of advantages if you didn't start with GraphQL in mind. It does an excellent job presenting ReST and RPC API as GraphQL. So you can quickly experiement with graph design if you're new to the concept. It also 'enhances' your graph with conventional queiry tools that make the api look more like SQL giving your filters and pagination out of the box. But if you've ever built a distributed system you're probably asking questions like:

  • How well does this perform?
  • What happens when the gateway fails?
  • How do the filters work?
  • How does it handle dependent system failures?

And of course, it handles them as well as it can, given this architecture: poorly.

Apollo Server is a developer tool that gets a poor-to-mid solution to market quickly. Because of that, its very attractive organisations that value time to market over performance. This trade off is nearly always the right solution.

But that's not why we're here. We want to do a native GraphQL solution. An actual GraphQL solution that is genuinely distributed, performant and fast to market.

This is how you do it.

Components

The mesh is composed of Repositories. These follow the Repository pattern documented by Martin Fowler in Patterns of Enterprise Architecture.

The premise is that we access data via a simple set of commands:

  • Create
  • CreateMany
  • Update
  • UpdateMany
  • Delete
  • DeleteMany

and then access the data on the repository:

  • Read
  • List

and finally allow us to query

  • getById
  • getByX

As these repositories exist in the network and not in memory we instantly fall foul of CAP Theory as we loose implicit control of state and expose ourselves to the far more likely network faults.

We mititate this by splitting the problem into two, similar to the CRQS pattern. Allowing us to scale commands and queries separately but, more importantly allowing us to use ReST for what it shines at: Commands and GraphQL for what it excels at Queries.

Restlettes

Responsible for event recording. They own the model, and are the only component allowed to write to the data store.

The provide a simple, and consistent way to store events. Simply define the swagger that describes the payload, and point them at a datastore.

You can use them for data retrieval, but just the state that you stored. No querying, no pulling data from other services.

Graphlettes

Responsible for turning a rag tag collection of restlettes and turning them into a mesh.

Each graphlette presents the world with its own view of the data starting from the data in its restlette:

N Restlettes == N Graphlettes

The are the primary means for extracting data from the system.

Microliths

Monoliths get a bad wrap. Given enough time they tend towards Big Ball of Mud We like to talk about S.O.L.I.D. and Object Orientated Programming as techniques to prevent this from happening, but the reality is that they, at best, delay this inevitability. The issue is less with how we build them, and more to do with how we use them. Success breeds more users, more users have more diverse demands, eventually the model on which the monolith is built breaks and the cost to make changes increases exponentially.

Microservices go in the other direction by moving objects into network. Surprisingly, its much easier to follow S.O.L.I.D. principals by building a series of black box services that agree on a common way of sharing information. But microservices push the complexity into the service and network layer.

We're taking a middle path. Composing monoliths with microservices.

We don't know your data or usage pattern. What we do know is that bundling multiple microservices in to a single service makes a lot of things easier.

Perfectly valid microliths:

  • A server per graphlette and restlette: true microservice
  • A single server with all graphlettes and restlettes: microlith
  • All graphlettes in one server, all restlettes in another
  • Low traffic xLettes in a microlith, high traffic xLettes in their own microservices.

We don't know what makes sense for you, but we have made it as easy as possible to change your mind.

We'd heavily recommend starting out with everything in a single microlith, then using data to figure out the best way to break it apart... and if you're wrong, merge them back together.

Its all very well having a means for an application to get data out of the system. But how can you inform other parts of the system that your data has changed?

Kafka, you use kafka.

This component attaches to your Mongo Replica Set and creates light weight events on public topics.

Other services are encouraged to consume those topics and then call the graph api to get the data they need to update their data.

Example

Applications are only allowed to present data that is owner by the user they represent.

In a simple case, the authorized_users is simply the user that created the event, but its not unusual for resources to be access by groups of users of certain roles.

So if we have a User service an Account service and a Group service that describes the Role and a list of User in that Group.

All services maintain a list of authorized_users but how users get into that list is an exercise for the system.

In this example, lets assume that we want all of the users in a group to be in the list of authorized_users.

To achieve this we create a consumer that listens to Groups topic, call the groups graph to figure out the current list of users, then modify the Accounts authorized_user list. Of course we may choose to listen to multiple groups, or even multiple services.

If an application needs to create an event, calling the restlettes makes the most sense. But what if the data is coming from inside of the enterprise?

In that instance we'd recommend creating an 'input topic' that keeps the data in same format as the restlette is expecting then using a Kafka Event Consumer to call the restlettes (single writer).

Only this component should read from the topic (it is private).

We don't know or care how events get into this topic, but we'd recommend something like:

System of Record => Change Data Capture => public topic => ksql => private topic => event consumer

Core Concepts

Modeling

Ownership

From the point of view of client applications, each repository owns the data and can be treated as a system of record.

It should be explicit which repository owns which data. Think 3rd normal form in RDBMS.

Complex objects are composed by meshing simple objects.

As we have learned from OO hierarchy can be an important tool, but is a tool of last resort. Composition is the only mechanism supported by gridql.

Advocacy

It is rarely the case that a Repository is actually the system of record. In our applications, the majority of the repositories mapped concepts owned by multiple external systems.

To that end it is the responsibility of the repository to mesh the demands of modern, transactional systems, with whatever has been patched together in the enterprise be they: flat files, rdbms, excel documents, manual entry.

Separation of Concerns and Awareness

Seperation Diagram

  • Clients must only be aware of the Ownership tier.
  • Repositories are only aware of each other and the queue
  • Advocates are aware of the queue, the Ownership tier and the Enterprise
    • But have really narrow purpose: create an event type, update a system of record
    • They are unaware of clients
  • Enterprise applications are unaware of everything

State

Idemptotency and Identification

As we are in a distributed multithreaded environment somethings that might be easy in a single threaded, monolithic solution become... interesting.

So we have three mechanisms of identification on a given object:

  1. Canonical "id" this field is automatically added to new objects in its meta data. Its a guid that is used to represent an object throughout its lifecycle in the ownership tier.
  2. The storage id "_id", also in metadata, but only used internally.
  • An updated object is a copy of the original with the same canonical id, but a different timestamp and _id
  • There can be N versions of an object in a timeseries.
  1. Payload id. These live in the payload and are used to tie an object back to an external system. They cannot be called "id"

Time

Because a query might happen whilst a repository is updating its collection, queries between repositories all contain the time stamp of the originating request.

This allows two features:

  1. Time travel: we can ask the system how it would have responded to the query at a given moment in the past
  2. Queries are consistent with the state of the mesh at the time of the query.

Asynchronisity

The ownership tier is distributed and synchronous. All commands and queries are responded to immediately.

However, reporting true state is important.

If you accept a change on behalf of a system of record you have a responsibility to make that truth available to client applications.

At least three patterns are effective here:

Hold state on the object
  1. Client Application submits a payload
  2. Repository emits an event
  3. An event propagator picks that up and writes a new field to the payload or meta data saying "received"
  4. Event propator propogates event to systems of record
  5. Event generator sees change in systems of record and writes back new state to object "accepted"
  6. Client application can query for new state

This mechanism is mostly useful when the object isn't part of a more complicated workflow.

One object per state
  1. Client Application submits a payload
  2. Repository emits an event
  3. Event propator propogates event to systems of record
  4. Event generator sees change in systems of record and creates an event that will generate a new object
  5. Client application can query for new object
Workflow
  1. Client Application submits a payload
  2. Repository emits an event
  3. Event propagator propagates event to systems of record
  4. Workflow sees repository event and updates its state for the transaction
  5. Event generator sees change in systems of record and creates an event that will generate a new object
  6. Workflow sees repository event for new object and updates its state for the transaction
  7. Client application can query the workflow for the state of the transaction and query appropriately

Authorization

Each repository is responsible for its own authorization. A basic 'auth' is provided that uses the SID on a JWT token, but fine grained authz support is encouraged.

Utilities

You went to the effort of creating json schema to describe your payloads... might as well use it to generate you test data.

Simple wrapper around json-schema to validate payloads.

You're probably better off using your own module.

Mongo does an excellent job of retrying if a connection is lost, but out of the box it won't retry an initial connection.

This is not a hard problem to solve, this is how we solved.

You're probably better off using your own module.

Example App

MicroBlog is a very quick an dirty 'Twitter Clone' to demonstrate how you can use this library for rapid role out of a data layer.

To Do

  • Graph Nodes
  • ReST Server
  • Mongo Listener that publishes to kafka
  • Bulk ReST API
  • Time aware GraphAPI
  • Auth
    • ReST
    • Graph
  • Kafka Listener that pushes to ReST

Coverage

coverage

gridql's People

Contributors

tsmarsh avatar tsmarshng avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.