Source originale du contenu
In December some of the tech guys at onefinestay invited me over to London to do some general improvements on their their nameko library. This collaboration came together because nameko was pretty similar to how I generally like to build certain infrastructure and I had some experience with very similar systems.
So now that some of those improvements hit the release version of nameko I figured it might be a good idea to give some feedback on why I like this sort of architecture.
Freeing your Mind
Right now if you want to build a web service in Python there are many tools you can pick from, but most of them live in a very specific part of your stack. The most common tool is a web framework and that will typically provide you with whatever glue is necessary to connect your own code to an incoming HTTP request that comes from your client.
However that's not all you need in an application. For instance very often you have periodic tasks that you need to execute and in that case, your framework is often not just not helping, it's also in your way. For instance because you might have built your code with the assumption that it has access to the HTTP request object. If you now want to run it from a cronjob that request object is unavailable.
In addition to crons there is often also the wish to execute something as the result of the request of a client, but without blocking that request. For instance imagine there is an admin panel in which you can trigger some very expensive data conversion task. What you actually want is for the current request to finish but the conversion task to keep on working in the background until your data set is converted.
There are obviously many existing solutions for that. Celery comes to mind. However they are typically very separated from the rest of the stack.
Having a system which treats all of this processes the same frees up your mind. This is what makes microservices interesting. Away with having HTTP request handlers that have no direct relationship with message queue worker tasks or cronjobs. Instead you can have a coherent system where any component can talk through well defined points with other parts of the system.
This is especially useful in Python where traditionally our support for parallel execution has been between very bad to abysmal.
Enter Nameko
Nameko is an implementation of this idea. It's very similar in architecture to how we structure code at Fireteam. It's based on distributing work between processes through AMQP. It's not just AMQP though. Nameko abstracts away from that and allows you to write your own transports, while staying true to the AMQP patterns.
Nameko does a handful of things and you can build very complex systems with it. The idea is that you build individual services which can emit events to which other services can subscribe to or they can directly invoke each other via RPC. All communication between the services happens through AMQP. You don't need to manually deal with any connectivity of those.
In addition to message exchange, services also use a lifecycle management to find useful resources through dependency injection. That sounds like a mouthful but is actually very simple. Because services are classes, you can add special attributes to them which will be resolved at runtime. The lifetime of the value resolved can be customized. For instance it becomes possible to attach a property to the class which can provide access to a database connection. The lifetime of that database connection can be automatically managed.
So how does that look in practice? Something like this:
from nameko.rpc import rpc
class HelloWorldService(object):
name = 'helloworld'
@rpc
def hello(self, name):
return "Hello, {}!".format(name)
This defines a basic service that provides one method that can be invoked via RPC. Either another service can do that, or any other process that runs nameko can also invoke that, for as long as they connect to the same AMQP server. To experiment with this service, Nameko provides a shell helper that launches an interactive Python shell with an n object that provides RPC access:
>>> n.rpc.helloworld.hello(name='John')
u'Hello, John!'
If the AMQP server is running, rpc.helloworld.hello contacts the helloworld service and resolves the hello method on it. Upon calling this method a message will be dispatched via the AMQP broker and be picked up by a nameko process. The shell will then block and wait for the result to come back.
A more useful example is what happens when services want to collaborate on some activity. For instance it's quite common that one service wants to respond to the changes another service performs to update it's own state. This can be achieved through events:
from nameko.events import EventDispatcher, event_handler
from nameko.rpc import rpc
class ServiceA(object):
name = 'servicea'
dispatch = EventDispatcher()
@rpc
def emit_an_event(self):
self.dispatch('my_event_type', 'payload')
class ServiceB(object):
name = 'serviceb'
@event_handler('servicea', 'my_event_type')
def handle_an_event(self, payload):
print 'service b received', payload
The default behavior is that one service instance of each service type will pick up the event. However nameko can also route an event to every single instance of every single service. This is useful for in-process cache invalidation for instance.
The Web
Nameko is not just good for internal communication however. It uses Werkzeug to provide a bridge to the outside world. This allows you to accept an HTTP request and to ingest a task into your service world:
import json
from nameko.web.handlers import http
from werkzeug.wrappers import Response
class HttpServiceService(object):
name = 'helloworld'
@http('GET', '/get/<int:value>')
def get_method(self, request, value):
return Response(json.dumps({'value': value}),
mimetype='application/json')
The endpoint function can itself invoke other parts of the system via RPC or other methods.
This functionality generally also extends into the websocket world, even though that part is currently quite experimental. It for instance is possible to listen to events and forward them into websocket connections.
Dependency Injection
One of the really neat design concepts in Nameko is the use of dependency injection to find resources. A good example is the SQLAlchemy bridge which attaches a SQLAlchemy database session to a service through dependency injection. The descriptor itself will hook into the lifecycle management to automatically manage the database resources:
from nameko_sqlalchemy import Session
import sqlalchemy as sa
from sqlalchemy.ext.declarative import declarative_base
Base = declarative_base()
class User(Base):
__tablename__ = 'users'
id = sa.Column(sa.Integer, primary_key=True)
username = sa.Column(sa.String)
class MyService(object):
name = 'myservice'
session = Session(Base)
@rpc
def get_username(self, user_id):
user = self.session.query(User).get(user_id)
if user is not None:
return user.username
The implementation of the Session dependency provider itself is ridiculously simple. The whole functionality could be implemented like this:
from weakref import WeakKeyDictionary
from nameko.extensions import DependencyProvider
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker
class Session(DependencyProvider):
def __init__(self, declarative_base):
self.declarative_base = declarative_base
self.sessions = WeakKeyDictionary()
def get_dependency(self, worker_ctx):
db_uri = self.container.config['DATABASE_URL']
engine = create_engine(db_uri)
session_cls = sessionmaker(bind=engine)
self.sessions[worker_ctx] = session = session_cls()
return session
def worker_teardown(self, worker_ctx):
sess = self.sessions.pop(worker_ctx, None)
if sess is not None:
sess.close()
The actual implementation is only a tiny bit more complicated, and that is basically just a bit of extra code to support different database URLs for different services and declarative bases. Overall the concept is the same however. When the dependency is needed, a connection to the database is established and when the worker shuts down, the session is closed.
Concurrency and Parallelism
What makes nameko interesting is that scales out really well through the use of AMQP and eventlet. First of all, when nameko starts a service container it uses eventlet to patch up the Python interpreter to support green concurrency. This allows a service container to become quite concurrent to do multiple things at once. This is very useful when a service waits on another service as threads in Python are a very disappointing story. As this however largely eliminates the possibility of true parallelism it becomes necessary to start multiple instances of services to scale up. Thanks to the use of AMQP however, this becomes a very transparent process. For as long as services do not need to store local state, it becomes very trivial to run as many of those service containers as necessary.
My Take On It
Nameko as it stands has all the right principles for building a platform out of small services and it's probably the best Open Source solution for this problem in the Python world so far.
It's a bit disappointing that Python's async story is so diverging between different Python versions and frameworks, but eventlet and gevent are by far the cleanest and most practical implementations, so for most intents and purposes the eventlet base in nameko is probably the best you can currently get for async IO. Fear not though, Nameko 2.0 now also runs on Python3.
If you haven't tried this sort of service setup yet, you might want to give Nameko a try.